On Apr 3, 2024, at 12:28 PM, Martin Bishop via cctalk
<cctalk(a)classiccmp.org> wrote:
Ignore my last - incontinence or is it incompetence
A fairly ordinary GPU, in a PC, could almost certainly provide an XY display with Z fade
(long persistance phosphor). I use them for waterfall displays and they keep up - the
data does of course arrive by E'net.
Yes, and some emulations have done this, such as Phil Budne's famous work in SIMH. I
adopted some of those ideas for DtCyber (CDC 6000 emulation) and it works well, though the
timing is marginal given the graphics subsystem I use. That is WxWidgets -- changes are
SDL would do better and some day I will try that.
Equally, FPGAs / SOCs can implement frame buffers; eg
to output waterfall displays. The fading memory would have to be in DRAM, FPGA memory is
fast but small 3 ns access time but only 240 ki by .. 2.18 Mi by (Zynq 10 .. 45, the
'45 is a corporate purchase). A ping pong buffer arrangement could implement fading -
computed in either processor (vector instructions) or logic (raw muscle). The DAC input
lines could supply the data.
Agreed, and that would be an elegant way to emulate a CDC DD60. Or a GT40. You'd
presumably want to use at least an HD level display signal (1920 by 1080), if not double
that, to make it look right; less than that would give pixel artefacts that make it not
look as vector-like as you would want.
paul