2019/9/24 13:00:21 -0700, [email protected]:
> Am 21.09.19 um 00:03 schrieb [email protected]:
>> To avoid this confusion, a more verbose specification might read:
>> * Returns the maximum number of $otype$s that will be produced for each
>> * $itype$ of input. This value may be used to compute the worst-case
>> size
>> * of the output buffer required for a given input sequence. This value
>> * accounts for any necessary content-independent prefix or suffix
>> #if[encoder]
>> * $otype$s, such as byte-order marks.
>> #end[encoder]
>> #if[decoder]
>> * $otype$s.
>> #end[decoder]
>
> wouldn't it be more clear to use "char" or even "{@code char}" instead
> "character" as replacment for the $xtype$ parameters?
The specifications of the Charset{De,En}coder classes make it clear
up front that “character” means “sixteen-bit Unicode character,” so
I don’t think changing “character” everywhere to “{@code char}” is
necessary.
This usage of “character” is common throughout the API specification.
With the introduction of 32-bit Unicode characters we started calling
those “code points,” but kept on calling sixteen-bit characters just
“characters.” (I don’t think the official term “Unicode code unit”
ever caught on, and it’s a bit of a mouthful anyway.)
- Mark