Why we are talking about some minor things like "umlauts" while personally I need to work with cyrillic symbols?
It is really strange to me this thread because I have no any troubles with multi-encoding text sources. May be I missed something. At the Linux UTF8 is natural and for the Windows we have nice LConvEncoding unit.
This is a good point. Many applications don't need to care about Unicode details at all. The details are encapsulated in library functions and GUI controls. For example iterating over Unicode characters is not needed very often really.
I also think the problems are exaggerated! Lots of energy wasted for the endless debates.
There are projects that must deal with Unicode details, for example Unicode aware editors like SynEdit, but they are a minority among all applications.
yeah true no one uses csv parsers, xml, json, xmpp, pascal, javascript, C or other language parsers, html enginees, rtf enginees, text drawing etc, only a small minority of projects do. Yeah right how about spam filters, search engines & sql filters?
Don't get me wrong I really don't care if the lcl/lazarus/fpc use only utf8, stay away from the basic types (ee unicodestring, ansistring, shortstring) that are supposed to have specific encoding/sizes (ee I couldn't care less what you do with the string type) so I don't have to rewrite, retest and reaprove my libraries and you can use what ever you want. Is it going to affect my future with lazarus? of course it is, speed, easy of use debugging capabilities, typing, everything is evaluated in a daily bases if you fall under a threshold I'm going to drop the tool.