That's different, and it's really just Nvidia saying that. Nvidia is pretending that each SIMD lane is a "core", even though they're tied together. By that definition, AMD's Bergamo 128-core EPYC would actually have 6144 "cores", because they can issue up to six 256-bit AVX instructions per clock.
Anyway, so no. These aren't like GPU "cores", they're real RISC-V CPU cores that can each execute a different program.
First, the term "bus" doesn't apply. I won't go into the details, but you can't do it with a bus. You need a packet-switched network of some kind.
Heh, that's cute. No, it's not as simple as you make it sound. Cache coherence is actually pretty challenging - especially as you try to scale it up.
en.wikipedia.org
Beyond that, you need both enough memory bandwidth and relatively low latency. Now, let's just look at the bandwidth aspect. The industry seems to have settled on having at least one 64-bit DDR5 DIMM per 8 cores. Since these are smaller, simpler cores, even a formula like 1 DIMM per 16 cores would mean having at least
100 DIMMs in the box with the 1600 core version.
Where are you going to put them all, and how are you going to correct them? If we consider that each of the latest EPYC CPUs can support up to 24 DIMMs, the 2-CPU boards with 48 DIMM slots hardly have room for anything else!
At SC22, we saw a Gigabyte GPU server with 48x DDR5 DIMM slots. We also saw a STH video on display and an Ampere Altra Arm with NVIDIA server
www.servethehome.com