We currently have two methods to detect unicode characters. Which is better? Primarily on performance. I want to consolidated. There might be other such code scattered!
public static boolean hasMultibyte(String value){ if( value == null )return false; for(int i = 0 ; i < value.length() ; i++ ){ char c = value.charAt(i); if(c > 0xFF )return true; } return false; } public static boolean isUnicodeFormat(final String format) { try { return !format.equals( new String(format.getBytes ("ISO-8859-1"), "ISO-8859-1")); } catch (UnsupportedEncodingException e) { return true; } } -- --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] Mailing List: http://jakarta.apache.org/site/mail2.html#poi The Apache Jakarta POI Project: http://jakarta.apache.org/poi/