[Haskell-cafe] Re: PROPOSAL: New efficient Unicode string library.

Stefan O'Rear stefanor at cox.net
Tue Oct 2 16:53:15 EDT 2007

On Tue, Oct 02, 2007 at 08:02:30AM -0700, Deborah Goldsmith wrote:
> UTF-16 is the type used in all the APIs. Everything else is considered an 
> encoding conversion.
> CoreFoundation uses UTF-16 internally except when the string fits entirely 
> in a single-byte legacy encoding like MacRoman or MacCyrillic. If any kind 
> of Unicode processing needs to be done to the string, it is first coerced 
> to UTF-16. If it weren't for backwards compatibility issues, I think we'd 
> use UTF-16 all the time as the machinery for switching encodings adds 
> complexity. I wouldn't advise it for a new library.

I do not believe that anyone was seriously advocating multiple blessed
encodings.  The main question is *which* encoding to bless.  99+% of
text I encounter is in US-ASCII, so I would favor UTF-8.  Why is UTF-16
better for me?

-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 189 bytes
Desc: Digital signature
Url : http://www.haskell.org/pipermail/haskell-cafe/attachments/20071002/81a95a11/attachment-0001.bin

More information about the Haskell-Cafe mailing list