Robert Haas wrote:
> On Sun, Mar 28, 2010 at 8:23 PM, Mike Rylander <mrylander@gmail.com> wrote:
>
>> In practice, every parser/serializer I've used (including the one I
>> helped write) allows (and, often, forces) any non-ASCII character to
>> be encoded as \u followed by a string of four hex digits.
>>
>
> Is it correct to say that the only feasible place where non-ASCII
> characters can be used is within string constants? If so, it might be
> reasonable to disallow characters with the high-bit set unless the
> server encoding is one of the flavors of Unicode of which the spec
> approves. I'm tempted to think that when the server encoding is
> Unicode we really ought to allow Unicode characters natively, because
> turning a long string of two-byte wide chars into a long string of
> six-byte wide chars sounds pretty evil from a performance point of
> view.
>
>
>
We support exactly one unicode encoding on the server side: utf8.
And the maximum possible size of a validly encoded unicode char in utf8
is 4 (and that's pretty rare, IIRC).
cheers
andrew