I think that one of the IMEs for Windows has a mode where it accepts hex
representing a Unicode code point - but I just tried to find this in the
Japanese IME without success.
- rick
-Original Message-
From: Patrick Andries [mailto:[EMAIL PROTECTED]
Sent: Friday, 14 November 2003 17:56
T
De: "Patrick Andries" <[EMAIL PROTECTED]>
> -Message d'origine -
> De: Murray Sargent
>
>
>
> >Patrick asks: «Q. How can I input any Unicode character if I know its
> hexadecimal code?»
> >
> >You could use an app that supports the Alt+x input method (like Word or
> WordPad)
> >and then co
-Message d'origine -
De: Murray Sargent
>Patrick asks: «Q. How can I input any Unicode character if I know its
hexadecimal code?»
>
>You could use an app that supports the Alt+x input method (like Word or
WordPad)
>and then copy the result into an app that doesn't.
[PA] No, this is no
John Cowan said:
> Kenneth Whistler scripsit:
>
> > However, there were character encoding standards committees,
> > predating the UTC, which did not understand this principle,
> > and which encoded a character for the Ångstrom sign as a
> > separate symbol. In most cases this would not be a prob
In a message dated 11/14/2003 2:34:26 PM Pacific Standard Time, [EMAIL PROTECTED] writes:
> what does i18n mean? I see it bandied about a lot.
It is a short hand for "IÃÅÃrÃÃÅÄÃÅÃÅÃÃn " because it is too hard for most of the people to type the "ÃÅÃrÃÃÅÄÃÅÃÅÃÃ" part. :) [and if your soft
John Cowan wrote:
> I need to write a normative reference to the
> Unicode Standard without specifying any particular
> version for inclusion in XML 1.1.
The W3C's Character Model (http://www.w3.org/TR/charmod) attempts to do this
as follows:
[Unicode]
The Unicode Consortium, The Unicode Stan
Jill Ramonsky asked on Nov 10:
> My question went unanswered, so I'll ask it again - do I get a vote?
Hmm... I'm finally catching up on mail-list mail from this past week. The
short answer to your question is "no", and others have said that. But, as
Philippe and others have said, you could jo
Patrick asks: «Q.
How can I input any Unicode character if I know its hexadecimal
code?»
You
could use an app that supports the Alt+x input method (like Word or WordPad) and
then copy the result into an app that doesn't.
For
reference, the Alt+x input method works as follows:
A hand
Kenneth Whistler scripsit:
> However, there were character encoding standards committees,
> predating the UTC, which did not understand this principle,
> and which encoded a character for the Ångstrom sign as a
> separate symbol. In most cases this would not be a problem,
> but in at least one Eas
Theodore H. Smith wrote:
Can someone give me some advice? If I was to write a dictionary class
for Unicode, would I be better off writing it using a b-tree, or
hash-bin system? Or maybe an array of pointers to arrays system?
See John's reply. Tries of some sort should be good. I think there was a
http://www.unicode.org/faq/font_keyboard.html states
:
«Q. How can I input any Unicode character if I know its hexadecimal
code?
A. Some platforms have methods of hexadecimal entry; others have only
decimal entry.On Windows, there is a decimal input method: hold down the
alt key while typi
Kenneth Whistler scripsit:
> My advice (maybe other editors will have a suggestion) would
> be something like picking a base version for reference and
> then specifying application of your reference to any subsequent
> versions defined by the Unicode Consortium. Thus you could
> say something like
Stefan Persson asked:
> Alexandre Arcouteil wrote:
> > Is that a clear indication that \u212B is actually a compatibility
> > character and then should be, according to XML 1.1 recommandation,
> > replaced by the \u00C5 character ?
>
> Isn't U+00C5 a compatibility character for U+0041 U+030A,
I need to write a normative reference to the Unicode Standard without
specifying any particular version for inclusion in XML 1.1. I can't
find anywhere on the Web site that explains how to do that, except
for just using the phrase "The Unicode Standard".
--
Values of beeta will give rise to dom!
internationalisation is just 20 chars long, hence the abbreviation.
Bernard
Theodore H. Smith wrote:
what does i18n mean? I see it bandied about a lot.
My guess is "internationalisation", but actually when you pronounce
"eye won ayht en" it doesn't sound anything like that word.
--
Theodor
Theodore H. Smith scripsit:
> Can someone give me some advice? If I was to write a dictionary class
> for Unicode, would I be better off writing it using a b-tree, or
> hash-bin system? Or maybe an array of pointers to arrays system?
Google for "ternary search trees". It's a very interesting
te
Ted Smith asked:
> what does i18n mean? I see it bandied about a lot.
>
> My guess is "internationalisation",
Correct. Or "internationalization", depending on your spelling
conventions.
> but actually when you pronounce
> "eye won ayht en" it doesn't sound anything like that word.
It is pron
Theodore H. Smith scripsit:
> what does i18n mean? I see it bandied about a lot.
>
> My guess is "internationalisation", but actually when you pronounce
> "eye won ayht en" it doesn't sound anything like that word.
Count the letters in "internationalisation" between the initial i and the
final n
From: "Theodore H. Smith" <[EMAIL PROTECTED]>
> what does i18n mean? I see it bandied about a lot.
>
> My guess is "internationalisation", but actually when you pronounce
> "eye won ayht en" it doesn't sound anything like that word.
Count the number of letters replaced by the two digits:
int
Try
a) ┐ etc.
b) Use an application to find those characters, copy them, and paste them into your HTML editor. For
this you need to use a Unicode charset for your HTML document, see
http://www.unicode.org/faq/unicode_web.html#9
Possible applications to use to find and copy the characters:
- Wi
From: "Kenneth Whistler" <[EMAIL PROTECTED]>
> This stuff *is* rather confusing for people encountering it the
> first time. But the above sources should help. Also see
> the W3C working draft for the Character Model for the World Wide Web
> 1.0:
>
> http://www.w3.org/TR/charmod/
You may need als
From: "John Cowan" <[EMAIL PROTECTED]>
> Philippe Verdy scripsit:
>
> > Generally, XML-based interfaces will perform normalization (NFC or NFD)
of
> > their input string, but they are not required to do it. However it
allows
> > the engine to guarantee that its outputs from canonically equivalent
s
Philippe Verdy scripsit:
> Generally, XML-based interfaces will perform normalization (NFC or NFD) of
> their input string, but they are not required to do it. However it allows
> the engine to guarantee that its outputs from canonically equivalent strings
> will also be canonically equivalent (be
From: "Kenneth Whistler" <[EMAIL PROTECTED]>
> Please disregard Philippe's misleading blatherings on this
> topic.
Thanks for denying all what I say, before finally saying the same thing as
me...
What I have said is not blathering, when I just say that UTR 20 is good only
in the context of text w
what does i18n mean? I see it bandied about a lot.
My guess is "internationalisation", but actually when you pronounce
"eye won ayht en" it doesn't sound anything like that word.
--
Theodore H. Smith - Macintosh Consultant / Contractor.
My website:
Can someone give me some advice? If I was to write a dictionary class
for Unicode, would I be better off writing it using a b-tree, or
hash-bin system? Or maybe an array of pointers to arrays system?
I suppose, that if I wanted an array of pointers to arrays, that I
couldn't use UTF32, I could
Joel,
I am posting your question to the Unicode list
www.unicode.org/consortium/distlist.html and hope that one of the
subscribers will have an answer to your question.
Magda Danish
Administrative Director
The Unicode Consortium
650-693-3010
[EMAIL PROTECTED]
-Original Message-
Date/Time
From: "Alexandre Arcouteil" <[EMAIL PROTECTED]>
> Philippe Verdy wrote:
>
> > From: "Kent Karlsson" <[EMAIL PROTECTED]>
> >
> >>Philippe Verdy wrote:
> >>
> >>>(1) a singleton (example the AngstrÃÂm symbol, canonically
> >>>mapped to A with diaeresis,
> >>
> >>The ÃângstrÃÂm (note spelling) si
Alexandre,
> Philippe Verdy wrote:
>
> > From: "Kent Karlsson" <[EMAIL PROTECTED]>
> >
> >>Philippe Verdy wrote:
> >>
> >>>(1) a singleton (example the Angström symbol, canonically
> >>>mapped to A with diaeresis,
> >>
> >>The Ångström (note spelling) sign is canonically mapped to
> >>cap
Alexandre Arcouteil wrote:
Is that a clear indication that \u212B is actually a compatibility
character and then should be, according to XML 1.1 recommandation,
replaced by the \u00C5 character ?
Isn't U+00C5 a compatibility character for U+0041 U+030A, so that both should be replaced by that?
S
Elaine Keown
still in Texas
RE: http://www.cox-internet.com/keownlaw/
Hi,
I'm sending this URL to ask for
FEEDBACK on FORMATTING.
I don't want to write more character descriptions
without getting some other opinions.
This list is NOT definitive--I had stuff lying ar
Philippe Verdy wrote:
From: "Kent Karlsson" <[EMAIL PROTECTED]>
Philippe Verdy wrote:
(1) a singleton (example the AngstrÃÂm symbol, canonically
mapped to A with diaeresis,
The ÃângstrÃÂm (note spelling) sign is canonically mapped to
capital a with ring.
Thanks for all explanations,
Keeping
Philippe Verdy posted:
I thought if had (it is used in French where it is clearly a typographic
ligature buf handled and sorted like two letters), as opposed to the ae
ligature (which is typographic ligature in French, but a true letter in
other
languages).
A strong case can be made that the _ae_
James Kass wrote:
In TrueType/OpenType, the first glyph in the font is used
as the "missing glyph".
And for Postscript fonts systems often take the bullet character in the
font as the "missing glyph" symbol.
However printers are very erratic in what they do.
Whether a character like ASCII 0x01
From: <[EMAIL PROTECTED]>
> Philippe Verdy wrote,
> The font specs strongly recommend that font developers use the
> narrow white box, or somthing very similar, for the missing glyph.
> But, some developers
Including Microsoft itself in some of its fonts installed with Windows XP...
Look at Symbol
From: "Kent Karlsson" <[EMAIL PROTECTED]>
> Philippe Verdy wrote:
> > (1) a singleton (example the AngstrÃÂm symbol, canonically
> > mapped to A with diaeresis,
> The ÃângstrÃÂm (note spelling) sign is canonically mapped to
> capital a with ring.
Beside the speeling (is it wrong to omit the ri
Philippe Verdy wrote:
>> Is somewhere a complete chart of "compatibility characters" ?
>
> Look at the Unicode data file which lists composition exclusions...
This is *by no means* a complete list of compatibility characters, in
either the strict or loose sense of the term.
-Doug Ewell
Fullert
Phillipe, instead of trying to sound authoritative by making up a whole-cloth
definition -- one that is completely and utterly wrong -- and thereby confuse
and mislead a beginner, you should either be silent or simply point the person
to the Unicode glossary:
http://www.unicode.org/glossary/#compa
.
Philippe Verdy wrote,
> But there are several fonts in Windows and Office that still display a
> normal question mark for this glyph ID, instead of a narrow white box as
> expected (this may be a caveat within the system compatibility font mappings
> with system fonts which are not TrueType but
Philippe Verdy wrote:
> (1) a singleton (example the AngstrÃm symbol, canonically
> mapped to A with diaeresis,
The ÃngstrÃm (note spelling) sign is canonically mapped to
capital a with ring.
There are several meanings of "compatibility characters".
The most important here are the charac
- Original Message -
From: "Alexandre Arcouteil" <[EMAIL PROTECTED]>
To: <[EMAIL PROTECTED]>
Sent: Friday, November 14, 2003 10:41 AM
Subject: compatibility characters (in XML context)
> This is a beginner question :
>
> In the XML 1.1 Proposed Recommendation 05 November 2003
> (http://w
> - -
> | User | | User |
> - -
> | App | | App |
> - -
> | Unicode | | Unicode |
> ---
> | Communication channel |
> ---
>
> In this model, Unicode ... Unicode offers as de
On 14/11/2003 00:20, Philippe Verdy wrote:
From: <[EMAIL PROTECTED]>
Please see
http://www.microsoft.com/typography/otspec/recom.htm
... the section about "Shape of .notdef glyph"
Thanks for pointing a Microsoft recommandation for the undefined glyph
(glyph id=0) that every TT font should
This is a beginner question :
In the XML 1.1 Proposed Recommendation 05 November 2003
(http://www.w3.org/TR/xml11), it is said that "Document authors are
encouraged to avoid "compatibility characters", as defined in section
6.8 of [Unicode]" so relating to Unicode 2.0.
I don't see any online d
From: "Jim Allan" <[EMAIL PROTECTED]>
> I take this to mean that any application can refuse to interpret PUA
> code points and still be conformant.
I would not say that, it would be excessive.
An application can use PUA the way it wants, but not line ill-formed
encoding sequences or non-characters
From: <[EMAIL PROTECTED]>
> Please see
> http://www.microsoft.com/typography/otspec/recom.htm
> ... the section about "Shape of .notdef glyph"
Thanks for pointing a Microsoft recommandation for the undefined glyph
(glyph id=0) that every TT font should implement (so this would affect also
OT fonts
Apologies for this cross-post.
All mail for me at either ccil.org or reutershealth.com between
6 PM and about 12 AM Thursday (2300 to 0500 UTC) for me was
unfortunately lost due to a botched spamassassin upgrade at ccil.org.
Please re-send any personal (non-mailing-list) mail to either
[EMAIL PROT
47 matches
Mail list logo