On Sat, 19 Feb 2011 23:43:33 +0000 Juan Jose Garcia-Ripoll <juanjose.garciarip...@googlemail.com> wrote:
> Would you find it useful to have an ECL that only supports character codes 0 > - 65535? That would make it probably easier to embed the part of the Unicode > database associated to it (< 65535 bytes) and have a standalone executable. > Executables would also be a bit faster and use less memory (16-bits vs > 32-bits per character) Would this be an option or would ECL internally use a 16-bit character representation all the time when unicode support is enabled for the build? Also I understand that the representation would take less memory but would it really be faster on 32-bit+ processors? I know that some processors (including older x86) have faster access times for 32-bit than 16-bit or 8-bit values (i.e. some time back I had to adapt an arcfour implementation to use 32-bit words rather than 8-bit ones for the internal state, despite it only holding values between 0 and 255, to enhance its performance). I also admit that I have some code assuming a 32-bit representation, but it's also ECL specific and could be adapted easily; I don't think that I make use of any character above 65535 myself. That said I have no idea what input I might have to deal with eventually, it's unpredictable. As for the 65535 bytes output file limitation, is that more difficult to fix? Is it a toolchain-dependent issue which ECL has no control over? Thanks, -- Matt ------------------------------------------------------------------------------ The ultimate all-in-one performance toolkit: Intel(R) Parallel Studio XE: Pinpoint memory and threading errors before they happen. Find and fix more than 250 security defects in the development cycle. Locate bottlenecks in serial and parallel code that limit performance. http://p.sf.net/sfu/intel-dev2devfeb _______________________________________________ Ecls-list mailing list Ecls-list@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ecls-list