On Tue, 30 Jun 2015, Dave Woyciesjes wrote:
I don't think this qualifies as answers persay,
but more just data
points really...
I have successfully installed & run Win7 x86 & x64 on Dell Latitude
D620, D630, D820 & D830. Not sure on the age, but they gotta be getting on to
around 7 years. The RAM they have varies between 2GB & 4GB.
I have also installed Win8 x64 on a Latitude D830, then proceeded to
swap that drive into a D620. Yesterday, I just "upgraded" a D820 from WIn7x64
to Win10 x64 preview; 3GB RAM, we'll see how that goes...
In other words, you should not be using WinXP anymore unless you have
an app that just won't work with Win7.
Why not??!?
Why do the experts advocate not using something that had been working?
The fact that you CAN "upgrade", doesn't seem to imply that you SHOULD.
In that case, ditch the program or run in a VM.
Why?
If the hardware is becoming too unreliable, . . .
If you need some sort of unavailable support, . . .
Otherwise, WHY change?