Jeremie Pelletier wrote:
Is there a way to convert a float (or double/real) to an integral number 
without changing its binary representation at compile time?

I need to extract the sign, exponent and mantissa yet I cant use bit shifting.
"Error: 'R' is not of integral type, it is a real" is the error I get.
There's a super-hacky way: pass the real as a template value parameter, and parse the .mangleof it! Not recommended, but it does work.

The other way is to do it with CTFE, subtracting powers of 2 until the residual is < 1.

Not great either.


The usual *cast(uint*)&value wont work either at compile time.

Any suggestions?

Reply via email to