The issue with your previous post isn't that you didn't type enough words. It's that you engaged in freewheeling speculation without bothering to check or cite any sources. If your goal is to impress people, you'll do better with one well-sourced claim than such pages of half-truths and backward-reasoning.
I was referring to how many dual-channel motherboards can support four DIMMs instead of just two by putting two DIMMs in each channel. This stresses the memory controller more than one DIMM per channel as two modules need to pull twice as much current (assuming identical DIMMs) and the lines going to both DIMM slots in one channel are even more difficult to make matched lengths and impedances than the individual lines going to a single DIMM slot.
Not sure sub-channels factors into the whole 2DPC issue. Seems a bit of a reach, to me.
Impedance-matching seems hard, when you have multiple devices per channel. DDR5 also features more training than DDR4 had, and it's not clear to me how much complexity it adds to actually
use the trained values when you have what is essentially a bus with multiple devices.
AMD's DDR5 memory controller often can't operate at the normal frequencies of DDR5, let alone higher frequencies with overclocking, when you have two DIMMs per channel
In case you haven't heard, AMD released new firmware that dramatically improved Ryzen 7000's memory speed and compatibility, even on 4-DIMM configurations.
Zen 4's memory controller is just as capable as Intel's — if not more so.
www.tomshardware.com
Prior generations of DDR, due simply to lower clock frequencies, were able to achieve much closer to full performance with two DIMMs per channel relative to one DIMM per channel than DDR5 can. With DDR3, the difference was negligible for people who weren't trying to get high memory overclocks.
My Sandybridge Xeon's CPU was only specified to run at DDR3-1333, if using 2 DIMMs per channel (unbuffered). At 1 DIMM per channel, it could do DDR3-1600. That would be as big of a drop as going from DDR5-5600 to DDR5-4550 (if such a speed existed). I didn't consider it negligible, at the time.
There was actually a PDF put out by Intel that detailed the speeds different memory configurations would run at. Another mitigating factor was using 1.35v instead of 1.5v memory, which I found out the hard way. That's what lead to me discovering the document.
PCIe went with the independent serial lanes instead of a variable width channel specifically for signal integrity. They could have made PCIe a parallel point-to-point connection capable of being 1 bit wide, 4 bits wide, 8 bits wide, or 16 bits wide if they wanted too,
You're completely ignoring bifurcation as a rationale for going with the multiple lanes approach. That gets way easier, if you have self-contained lanes.
I don't know which were the dominant factors in their design decisions, but I think neither do you. You might be right, but it only matters if we
know you're right, and that's something you only achieve by posting good (i.e. authoritative) sources.
PCIe being point-to-point wasn't entirely new either. Some boards had multiple PCI buses, sometimes individual PCI ports even had their own bus, which is essentially point-to-point then.
Implementations being point-to-point isn't the same as having a standard that's
fundamentally point-to-point. After waxing about multiple DIMMs per channel, I'm surprised you didn't immediately appreciate this.
In order to achieve those higher speeds, they had to use a narrower interface, ideally one as narrow as possible, which is serial for USB. Parallel ports being as slow as they were was why they could use such long cables.
You're getting history all muddled up. USB started out life at a max of 15 Mbps. The faster speeds only came along later. Serial wasn't needed to push past Parallel Ports' meager speed of 2 Mbps - USB opted to go with a serial bus for the other reasons mentioned.
They thought it would be completely impractical to try pushing IDE farther
Again, SCSI shows that's not true. UltraSCSI-320 is more than 2.5x as fast as the fastest PATA spec.
A key point you're glossing over with PATA is that it was a
bus standard, not unlike the original PCI. You could have
two devices per cable, which eventually should introduce issues like we saw with multiple DIMMs per channel.
However, it's not at all clear that being a parallel interconnect was a real limitation, at those speeds. It used a ribbon cable, with all the conductors being of the same length. However, it had the burden of legacy support, which might've been the biggest constraint holding it back.
They could have gone farther, but decided not to (except for the SAS folks who went one time farther than SATA) due to latency limitations of the SATA interface
SATA stopped at 6 Gbps because NVMe was just better in so many ways, rapidly maturing, and had already captured the high-end.
Like with USB, the smaller plug and cable were beneficial to commercial success but were only possible because of the serial interface in use
You have it backwards. The cost-savings and bulk reduction of SATA vs. PATA
drove the move from parallel to serial - not the other way around.
Again, the existence of DVI replacing VGA shows that a smaller connector was not needed to achieve commercial success for a new interface to replace an older one.
You're forgetting who bares the cost of PATA being such a clunky standard. It's disproportionately the OEMs. In contrast, it's overwhelmingly end-users dealing with video cables. OEMs drive the agenda, when it comes to PC standards.
UltraSCSI was a complicated and expensive interface. It had complicated and expensive hardware.
This was true back in the 1980's and early 1990's. Have you ever heard of ATAPI? It was used by many CD-ROM drives and was merely SCSI commands tunneled over IDE. Every controller supporting UDMA mode supported it.
Parallel interfaces as a whole were a product of the technology at the time. They could only make transistors switch efficiently so quickly at the time,
This is rubbish. If you go back to the origins of the parallel port cable, the reason they adopted it was to avoid having to add circuits to serialize and deserialize data over a serial link. In other words, the proliferation of parallel interfaces was much more about cost-savings than anything else.
en.wikipedia.org
Those were the pre-VLSI days. The main thing that changed is the cost-reductions and commodification of ICs needed to implement serial links.
DDR3 server boards, for example, often offered three registered DIMMs per channel with quad-channel boards often having twelve memory slots. I haven't seen any DDR5 boards do that.
DDR3-era CPUs had only up to 4 memory channels. DDR5-era server CPUs have either 8 or 12. With AMD, you get the same 12 DIMMs as those old 3DPC motherboards, but with 3x the bandwidth as 4-chanel/3DPC would provide. This lets you scale capacity along with bandwidth, making multiple DIMMs per channel is less important than it used to be.
Furthermore,
have you ever seen a 2-CPU server board with 24 DIMM slots?? It's hard to see where you'd fit
another 24! Here's one with the maximal 48-DIMM configuration - it's
huge and it has basically no room for anything else!
At SC22, we saw a Gigabyte GPU server with 48x DDR5 DIMM slots. We also saw a STH video on display and an Ampere Altra Arm with NVIDIA server
www.servethehome.com
Furthermore, in the modern era, die-stacking is reportedly set to usher in DDR5 DIMM capacities as large as 1 TB! At that point, you're spending so much $ on DRAM per system that it's probably not a big deal to buy some more CPUs, to scale up further.
Too many or modules, or substantial demand?
www.tomshardware.com
At today's prices 24 TB of 32 GB DDR5-4800 RDIMMs would cost a whopping $115,200. Obviously, GB/$ will improve by the time we have 1 TB DIMMs, but recent trends in semiconductor pricing suggest the improvement won't be much. Maybe a factor of 2 or 3? But, probably not the order of magnitude we're used to seeing.