On 01/09/2019 07:49 AM, Paul Koning via cctalk wrote:
Understanding rounding errors is perhaps the most
significant part of "numerical methods", a subdivision of
computer science not as widely known as it should be. I
remember learning of the work of a scientist at DEC whose
work was all about this: making the DEC math libraries not
only efficient but accurate to the last bit. Apparently
this isn't anywhere near as common as it should be. And I
wonder how many computer models are used for answering
important questions where the answers are significantly
affected by numerical errors. Do the authors of those
models know about these considerations? Maybe. Do the
users of those models know? Probably not. paul
A real problem on the IBM 360 and
370 was their floating
point scheme. They saved exponent bits by making the
exponent a power of 16, instead of 2. This meant that the
result of any calculation could end up normalized with up to
3 most-significant zeros. That would reduce the precision
of the number by up to 3 bits, or a factor of 8. Some
iterative solutions compared small differences in successive
calculations to decide when they had converged sufficiently
to stop.
These could either stop early, or run on for a long time
trying to reach convergence.
IBM eventually had to offer IEEE floating point format on
later machines.
Jon