On Tue, 2021-09-28 at 17:03 -0500, Jay Jaeger via cctalk wrote:
11:43 a.m., Vincent Long standing via cctalk wrote:
The C standards are more liberal, and continue to
to be 8 or more bits.
Was PL/I the only language that would let you select data
variables? Of course the fine print would not let you have more
decimal digits, or 32 bit binary. You would think by now that a
could handle any length data.
FORTRAN:? INTEGER*4 INTEGER*8 (and sometimes INTEGER*2 - e.g. Sun
FORTRAN-77)? was common, though never adopted as a standard.? Also
Fortran 90 introduced "kind type parameters" for all types. For REAL,
one can use SELECTED_REAL_KIND to ask for a specific number of decimal
digits. The standard does not require any minimum number be required.
Both "default real" and double precision are required. Many processors
provide quad precision. For INTEGER, one can use SELECTED_INT_KIND.
Processors are required to provide at least one kind with at least 18
decimal digits. There is no specification which other sizes are