How different is reading values from a `string` attribute compared to a 
`string` variable? If some software supports `string` variables shouldn't it 
support `string` attributes as well? If the CF is going to recommend `char` 
datatype for string-valued attributes, shouldn't the same be done for 
string-valued variables?

Prefixing the bytes of an UTF-8 encoded string with the BOM sequence is an odd 
practice. Although it is permitted, afaik, it is not recommended.

Since what gets stored are always the *bytes* of one string in some encoding, 
assuming UTF-8 always should take care of the ASCII character set, too. This 
could cause issues if someone used other one-byte encodings (e.g. ISO 8859 
family) but I don't see how such cases could be easily resolved.

Stroring Unicode strings using the `string` datatype makes more sense since the 
number of bytes for such strings is variable.

-- 
You are receiving this because you are subscribed to this thread.
Reply to this email directly or view it on GitHub:
https://github.com/cf-convention/cf-conventions/issues/141#issuecomment-407237484

Reply via email to