2019/9/24 13:00:21 -0700, ulf.zi...@cosoco.de: > Am 21.09.19 um 00:03 schrieb mark.reinh...@oracle.com: >> To avoid this confusion, a more verbose specification might read: >> * Returns the maximum number of $otype$s that will be produced for each >> * $itype$ of input. This value may be used to compute the worst-case >> size >> * of the output buffer required for a given input sequence. This value >> * accounts for any necessary content-independent prefix or suffix >> #if[encoder] >> * $otype$s, such as byte-order marks. >> #end[encoder] >> #if[decoder] >> * $otype$s. >> #end[decoder] > > wouldn't it be more clear to use "char" or even "{@code char}" instead > "character" as replacment for the $xtype$ parameters?
The specifications of the Charset{De,En}coder classes make it clear up front that “character” means “sixteen-bit Unicode character,” so I don’t think changing “character” everywhere to “{@code char}” is necessary. This usage of “character” is common throughout the API specification. With the introduction of 32-bit Unicode characters we started calling those “code points,” but kept on calling sixteen-bit characters just “characters.” (I don’t think the official term “Unicode code unit” ever caught on, and it’s a bit of a mouthful anyway.) - Mark