[Just a quick note to everyone that, I’ve just subscribed to this public list, 
and will look into this ongoing Mongolian-related discussion once I’ve mentally 
recovered from this week’s UTC stress. :)]

Best,
梁海 Liang Hai
https://lianghai.github.io

> On Jan 17, 2019, at 11:06, Asmus Freytag via Unicode <unicode@unicode.org> 
> wrote:
> 
> On 1/17/2019 9:35 AM, Marcel Schneider via Unicode wrote:
>>> [quoted mail]
>>> 
>>> But the French "espace fine insécable" was requested long long before 
>>> Mongolian was discussed for encodinc in the UCS. The problem is that the 
>>> initial rush for French was made in a period where Unicode and ISO were 
>>> competing and not in sync, so no agreement could be found, until there was 
>>> a decision to merge the efforts. Tge early rush was in ISO still not using 
>>> any character model but a glyph model, with little desire to support 
>>> multiple whitespaces; on the Unicode side, there was initially no desire to 
>>> encode all the languages and scripts, focusing initially only on trying to 
>>> unify the existing vendor character sets which were already implemented by 
>>> a limited set of proprietary vendor implementations (notably IBM, 
>>> Microsoft, HP, Digital) plus a few of the registered chrsets in IANA 
>>> including the existing ISO 8859-*, GBK, and some national standard or de 
>>> facto standards (Russia, Thailand, Japan, Korea).
>>> This early rush did not involve typographers (well there was Adobe at this 
>>> time but still using another unrelated technology). Font standards were 
>>> still not existing and were competing in incompatible ways, all was a mess 
>>> at that time, so publishers were still required to use proprietary software 
>>> solutions, with very low interoperability (at that time the only "standard" 
>>> was PostScript, not needing any character encoding at all, but only 
>>> encoding glyphs!)
>> 
>> Thank you for this insight. It is a still untold part of the history of 
>> Unicode.
> This historical summary does not square in key points with my own 
> recollection (I was there). I would therefore not rely on it as if gospel 
> truth.
> 
> In particular, one of the key technologies that brought industry partners to 
> cooperate around Unicode was font technology, in particular the development 
> of the TrueType Standard. I find it not credible that no typographers were 
> part of that project :).
> 
> Covering existing character sets (National, International and Industry) was 
> an (not "the") important goal at the time: such coverage was understood as a 
> necessary (although not sufficient) condition that would enable data 
> migration to Unicode as well as enable Unicode-based systems to process and 
> display non-Unicode data (by conversion). 
> 
> The statement: "there was initially no desire to encode all the languages and 
> scripts" is categorically false.
> 
> (Incidentally, Unicode does not "encode languages" - no character encoding 
> does).
> 
> What has some resemblance of truth is that the understanding of how best to 
> encode whitespace evolved over time. For a long time, there was a confusion 
> whether spaces of different width were simply digital representations of 
> various metal blanks used in hot metal typography to lay out text. As the 
> placement of these was largely handled by the typesetter, not the author, it 
> was felt that they would be better modeled by variable spacing applied 
> mechanically during layout, such as applying indents or justification.
> 
> Gradually it became better understood that there was a second use for these: 
> there are situations where some elements of running text have a gap of a 
> specific width between them, such as a figure space, which is better treated 
> like a character under authors or numeric formatting control than something 
> that gets automatically inserted during layout and rendering.
> 
> Other spaces were found best modeled with a minimal width, subject to 
> expansion during layout if needed.
> 
> 
> 
> There is a wide range of typographical quality in printed publication. The 
> late '70s and '80s saw many books published by direct photomechanical 
> reproduction of typescripts. These represent perhaps the bottom end of the 
> quality scale: they did not implement many fine typographical details and 
> their prevalence among technical literature may have impeded the 
> understanding of what character encoding support would be needed for true 
> fine typography. At the same time, Donald Knuth was refining TeX to restore 
> high quality digital typography, initially for mathematics.
> 
> However, TeX did not have an underlying character encoding; it was using a 
> completely different model mediating between source data and final output. 
> (And it did not know anything about typography for other writing systems).
> 
> Therefore, it is not surprising that it took a while and a few false starts 
> to get the encoding model correct for space characters.
> 
> Hopefully, well complete our understanding and resolve the remaining issues.
> 
> A./
> 
> 
> 
> 
> 
> 
> 

Reply via email to