Interesting question. My guess would have been two things:
- the chips are huge; capacity increase is more important than latency decrease
- caches are huge; spill and fill of cache lines is the important measure
- DRAM cells are getting smaller so the readout of a very tiny charge over a long distance is not getting easier
But then I found
this paper, which notes:
Quote:
In modern systems, DRAM-based main memory is significantly slower than the processor. Consequently, processors spend a long time waiting to access data from main memory, making the long main memory access latency one of the most critical bottlenecks in achieving high performance. Unfortunately, the latency of DRAM has remained almost constant in the past decade. The main reason for this is that DRAM is optimized for cost-per-bit (i.e., storage density), rather than access latency. Manufacturers leverage technology scaling to pack more DRAM cells in the same area, thereby enabling high DRAM density, as opposed to improving latency
So, it might be that a 1Mbit DRAM in today's technology would be very fast, but not worth making.
There's often a lot of memory on-chip these days, but AFAICT it's almost always SRAM, for speed and also because DRAM needs a different process compared to logic. But see
https://en.wikipedia.org/wiki/EDRAMfor in-module memory, which can be DRAM as it's a different chip.