What I never understood about this: how can big ENDian be explained ? Because, the number formats called big ENDian have the LEAST significant byte at the END, and the little ENDians have the MOST significant byte at the END.
Can anybody explain ? Regards Bernd Am Mit, 06 Aug 2003 schrieben Sie: > Has to do with the order of bytes and significance. Whether, for > example, decimal 123456789 which is hex 0x75BCD15 is stored as 07 5B CD > 15 (big endian) or 15 CD 5B 07 (little endian). Intel is little endian. > > ~ Daniel >
