Marvin,
I can always count on you for a detailed explanation. Thanks. You ought to turn
this into a blog post!
On Jun 17, 2010, at 4:06 PM, Marvin Humphrey wrote:
> There are two valid states for Perl scalars containing string data.
>
> * SVf_UTF8 flag off.
> * SVf_UTF8 flag on, and string data which is a valid UTF-8 byte sequence.
>
> In both cases, we define the logical content of the string as a series of
> Unicode code points.
>
> If the UTF8 flag is off, then the scalar's data will be interpreted as
> Latin-1. (Except under "use locale" but let's ignore that for now.) Each
> byte will be interpreted as a single code point. The 256 logical code points
> in Latin-1 are identical to the first 256 logical code points in Unicode.
> This is by design -- the Unicode consortium chose to overlap with Latin-1
> because it was so common. So any string content that consists solely of code
> points 255 and under can be represented in Latin-1 without loss.
Hrm. So am I safe in changing the CP1252 gremlin bytes to proper UTF-8
characters in Encode::ZapCP1252 like so?
$_[0] =~ s{([\x80-\x9f])}{
$table->{$1} ? Encode::decode('UTF-8', $table->{$1}) : $1
}emxsg
Where `$table` is the lookup table mapping hex values like \x80 to their UTF-8
equivalents (€)? This is assuming that $_[0] has the UTF8 flag on, of course.
So is this safe? Are \x80-\x9f considered characters when the utf8 flag is on,
or are they bytes that might break multibyte characters that use those bytes?
> In a Perl scalar with the UTF8 flag on, you can get the code points by
> decoding the variable width UTF-8 data, with each code point derived by
> reading 1-5 bytes. *Any* sequence of Unicode code points can be represented
> without loss.
Right.
> Unfortunately, it is really, really easy to mess up string handling when
> writing XS modules. A common error is to strip the UTF8 flag accidentally.
> This changes the scalar's logical content, as now its string data will be
> interpreted as Latin-1 rather than UTF-8.
>
> A less common error is to turn on the UTF8 flag for a scalar which does not
> contain a valid UTF-8 byte sequence. This puts the scalar into an what I'm
> calling an "invalid state". It will likely bring your program down with a
> "panic" error message if you try to do something like run a regex on it.
Fortunately, I'm not writing XS modules. :-)
> In your case, the Dump of the scalar demonstrated that it had the UTF8 flag
> set and that it contained a valid UTF-8 byte sequence -- a "valid state".
> However, it looks like it had invalid content.
Yes. I broke it with zap_cp1252 (applied before decoding). I just removed that
and things became valid again. The character was still broken, as it is in the
feed, but at least it was valid -- and the same as the source.
> A scalar with the UTF8 flag off can never be in an "invalid state", because
> any sequence of bytes is valid Latin-1. However, it's easy to change the
> string's logical content by accidentally stripping or forgetting to set the
> UTF8 flag. Unfortunately, this error leads to silent failure -- no error
> message, but the content changes -- and it can be really hard to debug.
Yes, this is what happened to me by zapping the non-utf8 scalar with zap_cp1252
before decoding it. Bad idea.
> This fellow's name, which you can see if you visit
> <http://twitter.com/tomaslau>, contains Unicode code point 0x010d, "LATIN
> SMALL
> LETTER C WITH CARON". As that code point is greater than 255, any Perl string
> containing his name *must* have the UTF8 flag turned on.
>
> I strongly suspect that at some point one of the following two things
> happened:
>
> * The code was input from a UTF-8 source but the input filehandle was not
> set to UTF-8 -- open (my $fh, '<:encoding(utf8)', $file) or die;
Well, I was pulling it from HTTP::Response->content. I'm not using
HTTP::Response->decoded_content because it's XML, which should be binary (see
http://juerd.nl/site.plp/perluniadvice)
> * The flag got stripped and subsequently the UTF-8 data was incorrectly
> reinterpreted as Latin-1.
> You typically need Devel::Peek for hunting down the second kind of error.
I missed that one, fortunately.
> It's more that getting UTF-8 support into Perl without breaking existing
> programs was a truly awesome hack -- but that one of the limitations of that
> hack was that the implementation is prone to silent failure.
Right. It's an impressive achievement. And I can't wait until DBI 2 is built on
Rakudo. ;-)
>> The output it still broken, however, in both cases, looking like this:
>>
>> Laurinavičius
>> LaurinaviÄius
>
> Let's double check something first. Based on your mail client (Apple Mail) I
> see you're (still) using OS X. Check out Terminal -> Preferences -> Advanced
> -> Character encoding. What's it set to? If it's not "Unicode (UTF-8)", set
> it to that now.
I always use UTF-8. Snow Leopard actually seems to allow multiple encodings
(!), as the "Encoding" tab (no more advanced tab) has UTF-8, Mac OS Roman,
Latin-1, and Latin-9 (wha?) checked, as well as a bunch of other encodings.
> Then try this:
>
> use 5.10.0;
> use Devel::Peek;
>
> my $str = "<p>Tomas Laurinavi\x{010d}ius</p>";
> say $str;
>
> binmode STDOUT, ':utf8';
> say $str;
>
> Dump $str;
> utf8::upgrade($str); # no effect
> Dump $str;
>
> For me, that prints his name correctly twice. The first time, though, I get
> a "wide character in print" warning. That warning arises because Perl's
> STDOUT is set to Latin-1 by default. It wants to "downgrade" the UTF8 scalar
> to Latin-1, but it can't do so without loss, so it warns and outputs the bytes
> as is. After we change STDOUT to 'utf8', the warning goes away.
Yep, same here.
> The utf8::upgrade() call has no effect, because the scalar starts off as
> UTF8. Prior to the introduction of the UTF8 flag, there was no way to put
> the code point \x{010d} into a Perl string because Latin-1 can't represent it.
> For backwards compatibility reasons, \x escapes below 255 have to be
> represented as Latin-1. Since you asked for \x{010d}, though, Perl knows
> that the backwards compat rules don't apply and it can use a UTF8 scalar.
Ah, I see. That's probably what happened inside Google Pipes: their code read
the original feed into a Latin-1 variable somehow, and the \x{010d} got changed
to \x{c4}\x{8d}, and it wasn't converted back before being output as UTF-8.
Best,
David