https://issues.dlang.org/show_bug.cgi?id=23573
Issue ID: 23573
Summary: std.bitmanip.bitfields doesn't respect native
endianness
Product: D
Version: D2
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: P1
Component: phobos
Assignee: [email protected]
Reporter: [email protected]
Given the following definition (-preview=bitfields)
```
union realbits
{
real value;
struct
{
version (LittleEndian)
{
private ubyte[6] padding;
ulong significand;
uint exp : 15;
bool sign : 1;
}
else
{
bool sign : 1;
uint exp : 15;
ulong significand;
private ubyte[6] padding;
}
}
}
```
You'd have thought that the equivalent std.bitmanip.bitfields would be
identical to the above, but no! You declare the bitfields in the same order,
regardless of the native endian in effect on the target.
```
import std.bitmanip : bitfields;
version (LittleEndian)
{
// ...
mixin(bitfields!(
uint, "exponent", 15,
bool, "sign" , 1));
}
else
{
version (THIS_GENERATES_WRONG_LAYOUT)
{
mixin(bitfields!(
bool, "sign" , 1,
uint, "exponent", 15));
}
else
{
mixin(bitfields!(
uint, "exponent", 15,
bool, "sign" , 1));
}
// ...
}
```
There could be a valid reason for this behaviour, and if so it should be
explicitly documented.
Otherwise, if this is indeed a bug, then it looks like the fix would be
something along the lines of foreach vs. foreach_reverse all bitfields
depending on little vs big endian respectively.
--