DO NOT REPLY TO THIS EMAIL, BUT PLEASE POST YOUR BUG RELATED COMMENTS THROUGH THE WEB INTERFACE AVAILABLE AT <http://nagoya.apache.org/bugzilla/show_bug.cgi?id=21491>. ANY REPLY MADE TO THIS MESSAGE WILL NOT BE COLLECTED AND INSERTED IN THE BUG DATABASE.
http://nagoya.apache.org/bugzilla/show_bug.cgi?id=21491 UTF-8 output is much slower for large chunks of character output ------- Additional Comments From [EMAIL PROTECTED] 2003-07-11 07:07 ------- In the case that the character, when converted into UTF-8 encoded bytes, may exceed the internal buffer the bytes are written one at a time directly to the OutputStream. This works, but the sheer volume of the number of calls is much slower than accumulating the bytes in an internal buffer. The "trick" in the patch that I am about to append is to logically break the character array up into chunks, each of which will not blow the output buffer, and for the write() method to call itself recursively with these chunks. This retains the buffering performance.
