On 19/03/2017 1:22 AM, Oleg B wrote:
Hello. I found strange behavior while casting enum array and immutable array.import std.stdio; void main() { enum arr = cast(ubyte[])[0,0,0,1,0,0,0,2,0,0,0,3,0,0,0,4]; auto arr1 = cast(void[])arr; immutable arr2 = cast(immutable(void)[])arr; enum arr3 = cast(void[])arr; writeln(cast(ushort[])arr1); // [0, 256, 0, 512, 0, 768, 0, 1024] writeln(cast(ushort[])arr2); // [0, 256, 0, 512, 0, 768, 0, 1024] writeln(cast(ushort[])arr3); // [0, 0, 0, 1, 0, 0, 0, 2, 0, 0, 0, 3, 0, 0, 0, 4] } I think it's related to representation of enums by compiler as #define. It's right? It's behavior by design?
It took me a bit but what I thinking happening is 1 and 2 are being casted at runtime where as 3 is at CT. At CT its per value of the enum array, at RT its per x bytes from array. Which sort of makes sense as at CT it does know the type even if you did cast it to void[].
