On 2015-06-01 02:46:53, Laszlo Ersek wrote:
> On 06/01/15 09:31, Jordan Justen wrote:
> > Supplementary Plane characters can exist in UTF-16 files,
> > but they are not valid UCS-2 characters.
> > 
> > For example, refer to this python interpreter code:
> >>>> import codecs
> >>>> codecs.encode(u'\U00010300', 'utf-16')
> > '\xff\xfe\x00\xd8\x00\xdf'
> > 
> > Therefore the UCS-4 0x00010300 character is encoded as two
> > 16-bit numbers (0xd800 0xdf00) in a little endian UTF-16
> > file.
> > 
> > For more information, see:
> > http://en.wikipedia.org/wiki/UTF-16#U.2B10000_to_U.2B10FFFF
> > 
> > This means that our current BaseTools code could be allowing
> > unsupported UTF-16 characters be used. To fix this, we decode the file
> > using python's utf-16 decode support. Then we verify that each
> > character's code point is 0xffff or less.
> > 
> > v3: Based on Mike Kinney's feedback, we now read the whole file and
> >     verify up-front that it contains valid UCS-2 characters. Thanks
> >     also to Laszlo Ersek for pointing out the Supplementary Plane
> >     characters.
> > 
> > Contributed-under: TianoCore Contribution Agreement 1.0
> > Signed-off-by: Jordan Justen <jordan.l.jus...@intel.com>
> > Cc: Yingke D Liu <yingke.d....@intel.com>
> > Cc: Michael D Kinney <michael.d.kin...@intel.com>
> > Cc: Laszlo Ersek <ler...@redhat.com>
> > ---
> >  BaseTools/Source/Python/AutoGen/UniClassObject.py | 84 
> > ++++++++++++++++++++++-
> >  1 file changed, 82 insertions(+), 2 deletions(-)
> > 
> > diff --git a/BaseTools/Source/Python/AutoGen/UniClassObject.py 
> > b/BaseTools/Source/Python/AutoGen/UniClassObject.py
> > index aa54f4f..66fdbf0 100644
> > --- a/BaseTools/Source/Python/AutoGen/UniClassObject.py
> > +++ b/BaseTools/Source/Python/AutoGen/UniClassObject.py
> > @@ -19,6 +19,7 @@
> >  import Common.LongFilePathOs as os, codecs, re
> >  import distutils.util
> >  import Common.EdkLogger as EdkLogger
> > +import StringIO
> >  from Common.BuildToolError import *
> >  from Common.String import GetLineNo
> >  from Common.Misc import PathClass
> > @@ -147,6 +148,33 @@ def GetLanguageCode(LangName, IsCompatibleMode, File):
> >  
> >      EdkLogger.error("Unicode File Parser", FORMAT_INVALID, "Invalid RFC 
> > 4646 language code : %s" % LangName, File)
> >  
> > +## Ucs2Codec
> > +#
> > +# This is only a partial codec implementation. It only supports
> > +# encoding, and is primarily used to check that all the characters are
> > +# valid for UCS-2.
> > +#
> > +class Ucs2Codec(codecs.Codec):
> > +    def __init__(self):
> > +        self.__utf16 = codecs.lookup('utf-16')
> > +
> > +    def encode(self, input, errors='strict'):
> > +        for Char in input:
> > +            if ord(Char) > 0xffff:
> > +                raise ValueError("Code Point too large to encode in UCS-2")
> > +        return self.__utf16.encode(input)
> 
> I could be missing some context, but the wikipedia article referenced at
> the top says that
> 
>   The official Unicode standard says that no UTF forms [...] can encode
>   [U+D800 to U+DFFF] code points.
> 
>   However UCS-2, UTF-8 [...] can encode [U+D800 to U+DFFF] code points
>   in trivial and obvious ways [...]
> 
> So basically it is possible to construct a UTF-8 file that (albeit
> violating the unicode standard) will result in U+D800 to U+DFFF code
> *points* when parsed. Then (I think) these would be turned into UCS2
> code *units* with identical numeric values. Which, I think, would be wrong.
> 
> Also, the same seems to apply to a UTF-16 source:
> 
>   It is possible to unambiguously encode them in UTF-16 by using a code
>   unit equal to the code point, as long as no sequence of two code
>   units can be interpreted as a legal surrogate pair (that is, as long
>   as a high surrogate is never followed by a low surrogate). The
>   majority of UTF-16 encoder and decoder implementations translate
>   between encodings as though this were the case.
> 
> Shouldn't we accept code points *only* in "U+0000 to U+D7FF and U+E000
> to U+FFFF"? The article says

I think you are right. Under "U+D800 to U+DFFF" it says: "The Unicode
standard permanently reserves these code point values for UTF-16
encoding of the high and low surrogates, and they will never be
assigned a character, so there should be no reason to encode them."

So, I guess there is no valid reason to accept them as input code
points.

-Jordan

>     Both UTF-16 and UCS-2 encode code points in this range as single
>     16-bit code units that are numerically equal to the corresponding
>     code points. These code points in the BMP are the /only/ code points
>     that can be represented in UCS-2. Modern text almost exclusively
>     consists of these code points.
> 
> For example,
> 
>   printf '%b' '\x01\xD8\x24\x00'
> 
> creates two UTF-16 code units (in LE representation):
> 
>   D801 0024
> 
> This is not a valid surrogate pair (there is no low surrogate). Does
> your code (or the underlying python machinery) reject the output of the
> above printf invocation?
> 
> For example, "iconv" rejects it:
> 
> $ printf '%b' '\x01\xD8\x24\x00' | iconv -f UTF-16LE -t UCS-2
> iconv: illegal input sequence at position 0
> 
> Hm... let's see a bit lower down:
> 
> > +
> > +TheUcs2Codec = Ucs2Codec()
> > +def Ucs2Search(name):
> > +    if name == 'ucs-2':
> > +        return codecs.CodecInfo(
> > +            name=name,
> > +            encode=TheUcs2Codec.encode,
> > +            decode=TheUcs2Codec.decode)
> > +    else:
> > +        return None
> > +codecs.register(Ucs2Search)
> > +
> >  ## StringDefClassObject
> >  #
> >  # A structure for language definition
> > @@ -209,7 +237,7 @@ class UniFileClassObject(object):
> >          Lang = distutils.util.split_quoted((Line.split(u"//")[0]))
> >          if len(Lang) != 3:
> >              try:
> > -                FileIn = codecs.open(LongFilePath(File.Path), mode='rb', 
> > encoding='utf-16').read()
> > +                FileIn = self.OpenUniFile(LongFilePath(File.Path))
> >              except UnicodeError, X:
> >                  EdkLogger.error("build", FILE_READ_FAILURE, "File read 
> > failure: %s" % str(X), ExtraData=File);
> >              except:
> > @@ -253,6 +281,58 @@ class UniFileClassObject(object):
> >                      self.OrderedStringDict[LangName][Item.StringName] = 
> > len(self.OrderedStringList[LangName]) - 1
> >          return True
> >  
> > +    def OpenUniFile(self, FileName):
> > +        #
> > +        # Read file
> > +        #
> > +        try:
> > +            UniFile = open(FileName, mode='rb')
> > +            FileIn = UniFile.read()
> > +            UniFile.close()
> > +        except:
> > +            EdkLogger.Error("build", FILE_OPEN_FAILURE, ExtraData=File)
> > +
> > +        #
> > +        # We currently only support UTF-16
> > +        #
> > +        Encoding = 'utf-16'
> > +
> > +        self.VerifyUcs2Data(FileIn, FileName, Encoding)
> > +
> > +        UniFile = StringIO.StringIO(FileIn)
> > +        Info = codecs.lookup(Encoding)
> > +        (Reader, Writer) = (Info.streamreader, Info.streamwriter)
> > +        return codecs.StreamReaderWriter(UniFile, Reader, Writer)
> > +
> > +    def VerifyUcs2Data(self, FileIn, FileName, Encoding):
> > +        Ucs2Info = codecs.lookup('ucs-2')
> > +        #
> > +        # Convert to unicode
> > +        #
> > +        try:
> > +            FileDecoded = codecs.decode(FileIn, Encoding)
> 
> If this method call is guaranteed to throw an exception when it
> encounters a UTF-16LE (or UTF-8) sequence that would decode to a
> surrogate code point (ie. in the U+D800 to U+DFFF range), then we should
> be fine.
> 
> Otherwise, such code points should be caught in Ucs2Codec.encode().
> 
> Thanks
> Laszlo
> 
> 
> > +            Ucs2Info.encode(FileDecoded)
> > +        except:
> > +            UniFile = StringIO.StringIO(FileIn)
> > +            Info = codecs.lookup(Encoding)
> > +            (Reader, Writer) = (Info.streamreader, Info.streamwriter)
> > +            File = codecs.StreamReaderWriter(UniFile, Reader, Writer)
> > +            LineNumber = 0
> > +            ErrMsg = lambda Encoding, LineNumber: \
> > +                     '%s contains invalid %s characters on line %d.' % \
> > +                     (FileName, Encoding, LineNumber)
> > +            while True:
> > +                LineNumber = LineNumber + 1
> > +                try:
> > +                    Line = File.readline()
> > +                    if Line == '':
> > +                        EdkLogger.error('Unicode File Parser', 
> > PARSER_ERROR,
> > +                                        ErrMsg(Encoding, LineNumber))
> > +                    Ucs2Info.encode(Line)
> > +                except:
> > +                    EdkLogger.error('Unicode File Parser', PARSER_ERROR,
> > +                                    ErrMsg('UCS-2', LineNumber))
> > +
> >      #
> >      # Get String name and value
> >      #
> > @@ -305,7 +385,7 @@ class UniFileClassObject(object):
> >              EdkLogger.error("Unicode File Parser", FILE_NOT_FOUND, 
> > ExtraData=File.Path)
> >  
> >          try:
> > -            FileIn = codecs.open(LongFilePath(File.Path), mode='rb', 
> > encoding='utf-16')
> > +            FileIn = self.OpenUniFile(LongFilePath(File.Path))
> >          except UnicodeError, X:
> >              EdkLogger.error("build", FILE_READ_FAILURE, "File read 
> > failure: %s" % str(X), ExtraData=File.Path);
> >          except:
> > 
> 

------------------------------------------------------------------------------
_______________________________________________
edk2-devel mailing list
edk2-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/edk2-devel

Reply via email to