开发者

What is the rationale for exponent and mantissa sizes in IEEE floating point standards?

I have a decent underst开发者_运维问答anding of how floating point works, but I want to know how the specific exponent and mantissa sizes were decided upon. Are they optimal in some way? How can optimality be measured for floating point representations (I assume there are several ways)? I imagine these issues are addressed in the official standard, but I don't have access to it.


According to this interview with Will Kahan, they were based on the VAX F and G formats of the era.

Of course that doesn't answer the question of how those formats were chosen...


For 32-bit IEEE floats, the reasoning is that the precision should be at least as good as 24 bits fixed point.

Why exactly 24 bits, I don't know, but it seems like a reasonable tradeoff.

I suppose having a nice "round" number like that (mantissa + sign = 3 bytes, exponent = 1 byte) can also make implementations more efficient.

0

上一篇:

下一篇:

精彩评论

暂无评论...
验证码 换一张
取 消

最新问答

问答排行榜