On September 12, 2013 Richard Wordingham wrote:
On Wed, 11 Sep 2013 12:57:34 +0200
Charlie Ruland ☘ <[email protected]> wrote:
Andreas,
linguistically speaking (i.e. following the tradition that was
started by Ferdinand de Saussure) when items are used contrastively
they must be considered different linguistic entities on what has
been called the “emic” level: phonemes, morphemes, graphemes, etc.
But this switching is not done at the character level, but at the word
level.  To follow the linguistic comparison, this is code-switching.  If
an utterance contains a mix of English and French, one doesn't normally
use one character for English /s/ and another for French /s/ in
broad phonetic transcription ('emic' level). One might choose different
fonts, or font effects, to mark which language was being used, but that
is a different matter.  One might simply colour code the difference, but
that would be regarded as extravagant.
That’s correct, script switching is not done at the segmental, or character, level. And I surely didn’t meant to say that a new set of characters would need to be encoded. I for one would have been happy with a unification of all the Brahmi based scripts of India, as had been the attitude of the Indian government at one time; nor do I object to the unification of, say, Japanese 禅 and simplified Chinese 禅 (U+7985; I have used the locale tags ja and zh-Hans, hopefully you can see the difference — cf. U+5358 単 and U+5355 单, which are not unified). Dealing with contrasts in writing system doesn’t necessarily imply encoding different characters, but finding a way to show these contrasts on whichever level is appropriate. (Incidentally this is to some extent paralleled in the realm of phonology, where the notion of segmental phonemes is not exactly helpful when it comes to stress, intonation etc.)

As to English /s/ and French /s/: No matter if their symbolic representations look the same or not one must always bear in mind that the two /s/’s have quite different systemic values and thus refer to clearly different things. I am sure you are aware of this.
At a practical level, when writing about a computer program, I find it
useful to use a proportional font for normal text and a monospaced
text for identifiers within the program, but I wouldn't dream of
claiming that the identifiers were really written in a different script
to the plain English in the discussion.
For speakers of languages other than English the inventory of characters used may be quite different: for example, ⟨ß⟩ ⟨ơ⟩ are used in “normal” text in German and Vietnamese respectively, whereas ⟨@⟩ ⟨#⟩ and in Vietnamese also ⟨f⟩ ⟨j⟩ ⟨w⟩ ⟨z⟩ are usually confined to the sphere of computers. But this is a matter of alphabet rather than script.
As /gebrochene Schrift/ and /Antiqua/ were habitually used
contrastively there is no doubt that they are different scripts in
that tradition, although they may be the same script in another
tradition. This is very much like [ɛ] and [æ] being different
linguistic entities (phonemes) in English, but not in German.
It's not so very different to using different fonts to indicate
different languages when the text is actually 'translated' for the
readers convenience, e.g. the use of an Arabic-looking font to
represent Klatchian in the Discworld novel 'Jingo'.  I wouldn't suggest
that we have a different script for the Klatchian; rather, the novel is
not in plain text in these sections.
A solution that is more general that specifying fonts which may or may not be installed on a system is to use different locale tags for “blackletter Anne” from Berlin and “antiqua Anne” from Paris: Anne (which I have tagged de-Latf) vs. Anne (de-Latn). If you can’t see a difference here it is because your e-mail viewing software doesn’t use different fonts for display. IMO the W3C should be prompted to stipulate that distinguishable letters be used for spans of text that are script-tagged Latn, Latf and Latg. Though this is beyond the scope of Unicode each of the two standards depends on the other. /Tout se tient./

One final remark: Thinking about it I have the impression that the blackletter vs. antiqua distinction once made in German very much resembles that made between Hiragana and Katakana in Japanese. In both cases the underlying systems of the corresponding scripts are essentially the same; yet it seems impossible to read the other script without further instruction and exercise; and in both languages one script is used primarily for inherited, and the other for foreign words.

Charlie
Richard.

Reply via email to