On 02/20/2015 09:47 PM, Eric Smith wrote:
One can argue that a not-quite-C compiler for a PDP-10
that supported
6-, 7-, and 8-bit character types as well as 18-bit and 36-bit
integers would be useful, but by definition it wouldn't actually be C.
What does the C standard say about the physical size of an int in bits,
versus the largest and smallest integer that can be held in those bits?
Consider a 64 bit int which is a simple degenerate floating point number
with 16 bits of mantissa (presumably set to 0) and 48 signed bits of
mantissa. All math done in unnormalized floating point.
Does C require 1's or 2's complement or can it support sign-magnitude
integers?
--Chuck