On Tue, 2021-09-28 at 17:03 -0500, Jay Jaeger via cctalk wrote:
On 2021-09-28
11:43 a.m., Vincent Long standing via cctalk wrote:
The C standards are more liberal, and continue to
require char
types
to be 8 or more bits.
Was PL/I the only language that would let you select data
size for
variables? Of course the fine print would not let you have more
than 16
decimal digits, or 32 bit binary. You would think by now that a
language
could handle any length data.
Hardly.
FORTRAN:? INTEGER*4 INTEGER*8 (and sometimes INTEGER*2 - e.g. Sun
FORTRAN-77)? was common, though never adopted as a standard.? Also
REAL
vs. DOUBLE.
Fortran 90 introduced "kind type parameters" for all types. For REAL,
one can use SELECTED_REAL_KIND to ask for a specific number of decimal
digits. The standard does not require any minimum number be required.
Both "default real" and double precision are required. Many processors
provide quad precision. For INTEGER, one can use SELECTED_INT_KIND.
Processors are required to provide at least one kind with at least 18
decimal digits. There is no specification which other sizes are
required.