What is the rationale for exponent and mantissa sizes in IEEE floating point standards?

蹲街弑〆低调 提交于 2019-11-27 08:23:08

问题


I have a decent understanding of how floating point works, but I want to know how the specific exponent and mantissa sizes were decided upon. Are they optimal in some way? How can optimality be measured for floating point representations (I assume there are several ways)? I imagine these issues are addressed in the official standard, but I don't have access to it.


回答1:


According to this interview with Will Kahan, they were based on the VAX F and G formats of the era.

Of course that doesn't answer the question of how those formats were chosen...




回答2:


For 32-bit IEEE floats, the reasoning is that the precision should be at least as good as 24 bits fixed point.

Why exactly 24 bits, I don't know, but it seems like a reasonable tradeoff.

I suppose having a nice "round" number like that (mantissa + sign = 3 bytes, exponent = 1 byte) can also make implementations more efficient.



来源:https://stackoverflow.com/questions/4397081/what-is-the-rationale-for-exponent-and-mantissa-sizes-in-ieee-floating-point-sta

易学教程内所有资源均来自网络或用户发布的内容,如有违反法律规定的内容欢迎反馈
该文章没有解决你所遇到的问题?点击提问,说说你的问题,让更多的人一起探讨吧!