Andrew Dunstan wrote:
>
>
> Robert Haas wrote:
>> On Sun, Mar 28, 2010 at 8:23 PM, Mike Rylander <mrylander@gmail.com>
>> wrote:
>>
>>> In practice, every parser/serializer I've used (including the one I
>>> helped write) allows (and, often, forces) any non-ASCII character to
>>> be encoded as \u followed by a string of four hex digits.
>>>
>>
>> Is it correct to say that the only feasible place where non-ASCII
>> characters can be used is within string constants? If so, it might be
>> reasonable to disallow characters with the high-bit set unless the
>> server encoding is one of the flavors of Unicode of which the spec
>> approves. I'm tempted to think that when the server encoding is
>> Unicode we really ought to allow Unicode characters natively, because
>> turning a long string of two-byte wide chars into a long string of
>> six-byte wide chars sounds pretty evil from a performance point of
>> view.
>>
>>
>>
>
> We support exactly one unicode encoding on the server side: utf8.
>
> And the maximum possible size of a validly encoded unicode char in
> utf8 is 4 (and that's pretty rare, IIRC).
>
>
Sorry. Disregard this. I see what you mean.
Yeah, I thing *requiring* non-ascii character to be escaped would be evil.
cheers
andrew