Can someone explain to me how I convert a 32-bit floating point value to a 16-bit floating point value?
(s = sign e = exponent and m = mantissa)
If 32-bit float is 1s7e24m
And 16-bit float is 1s5e10m
Then is it as simple as doing?
int fltInt32; short fltInt16; memcpy( &fltInt32, &flt, sizeof( float ) ); fltInt16 = (fltInt32 & 0x00FFFFFF) >> 14; fltInt16 |= ((fltInt32 & 0x7f000000) >> 26) << 10; fltInt16 |= ((fltInt32 & 0x80000000) >> 16);
I'm assuming it ISN'T that simple ... so can anyone tell me what you DO need to do?
Edit: I cam see I've got my exponent shift wrong ... so would THIS be better?
fltInt16 = (fltInt32 & 0x007FFFFF) >> 13; fltInt16 |= (fltInt32 & 0x7c000000) >> 13; fltInt16 |= (fltInt32 & 0x80000000) >> 16;
I'm hoping this is correct. Apologies if I'm missing something obvious that has been said. Its almost midnight on a friday night ... so I'm not "entirely" sober ;)
Edit 2: Ooops. Buggered it again. I want to lose the top 3 bits not the lower! So how about this:
fltInt16 = (fltInt32 & 0x007FFFFF) >> 13; fltInt16 |= (fltInt32 & 0x0f800000) >> 13; fltInt16 |= (fltInt32 & 0x80000000) >> 16;
Final code should be:
fltInt16 = ((fltInt32 & 0x7fffffff) >> 13) - (0x38000000 >> 13); fltInt16 |= ((fltInt32 & 0x80000000) >> 16);
The exponent needs to be unbiased, clamped and rebiased. This is the fast code I use:
unsigned int fltInt32; unsigned short fltInt16; fltInt16 = (fltInt32 >> 31) << 5; unsigned short tmp = (fltInt32 >> 23) & 0xff; tmp = (tmp - 0x70) & ((unsigned int)((int)(0x70 - tmp) >> 4) >> 27); fltInt16 = (fltInt16 | tmp) << 10; fltInt16 |= (fltInt32 >> 13) & 0x3ff;
This code will be even faster with a lookup table for the exponent, but I use this one because it is easily adapted to a SIMD workflow.
Limitations of the implementation:
2^-14instead of zero.
Be careful with denormals. If your architecture uses them, they may slow down your program tremendously.
The exponents in your float32 and float16 representations are probably biased, and biased differently. You need to unbias the exponent you got from the float32 representation to get the actual exponent, and then to bias it for the float16 representation.
Apart from this detail, I do think it's as simple as that, but I still get surprised by floating-point representations from time to time.
Check for overflow when doing the thing with the exponents while you're at it.
Your algorithm truncates the last bits of the mantisa a little abruptly, that may be acceptable but you may want to implement, say, round-to-nearest by looking at the bits that are about to be discarded. "0..." -> round down, "100..001..." -> round up, "100..00" -> round to even.
Here's the link to an article on IEEE754, which gives the bit layouts and biases.
User contributions licensed under CC BY-SA 3.0