On Mon, 20 Sep 2010 20:51:20 +0200 <[email protected]> wrote: > Hello > > I had some issues with LConvEncoding as I wrote in bugtracker issue #0017212 > > My mistake was essentially the meaning of character encoding "ANSI". Sorry > for that. > > So I took a deeper view at this unit. It is far not complete concerning > existing codepages and support for Unicode.
It's goal is to convert the most common codepages and do the rest with the existing OS libraries (iconv). A complete conversion unit should be part of FPC. > Further, the implementation with string-based encoding-descriptions is > not optimal regarding speed and the possibilty to extend the unit. The > encoding and decoding of asian-codepages is really suboptimal. (a fast > and dirty performance check showed that it is about 3000 times slower > than singlebyte codepages like ISO-8859-1) True. And the tables are quite big. Maybe the unit should be splitted into several units. > I would rewrite the entire unit, including most of the codepages from > ftp.unicode.org, full support for UTF-8, UTF-16 and UTF-32, including LE and > BE variants. And as a further utility, a conversion program for the mentioned > codepagefiles, that will create a complete pascal include file with all the > necessary tables and conversion functions. What do you mean with full support for UTF-8, UTF-16 and UTF-32? Mattias -- _______________________________________________ Lazarus mailing list [email protected] http://lists.lazarus.freepascal.org/mailman/listinfo/lazarus
