hatib :
bit_user :
hatib :
guys imagine 24tb memory 8 xeon e7 8890 v4 and 24 tesla p100 that would be awesome
The question is: what would you do with such a thing?
for video editing and simulation
By video editing, perhaps you have in mind some complicated motion graphics effects and rendering? In any case, I think just a single-CPU box with a lot of RAM, fast storage, and a fast GPU would be enough for everything in video editing/production, except big rendering jobs.
It's actually a little tricky to think of tasks that would benefit from having all this in one box, but not suffer from the NUMA topology. Keep in mind that the memory & PCIe busses are partitioned amongst the CPUs, and that accessing a device or memory attached to another CPU means going through QPI. So, that would become a bottleneck, for certain workloads.
These systems have a sweet spot. For a task to run well on these systems, it would need to be latency and/or bandwidth sensitive beyond the limits of 10/40/100 gigabit ethernet, but not so much that QPI would become a bottleneck.
Many simulation & server tasks would perform similarly on cheaper, single-CPU boxes networked with 10/40/100 gigabit, which has the added benefit of scaling better. There's a nice writeup on nextplatform.com of how Facebook migrated from dual-E5 Xeons to single-CPU Xeon D systems, for better efficiency. For them, it seems QPI was just an overhead (in terms of power), since they had separable tasks which didn't need a lot of intercommunication.
That's why I go to realtime (i.e. highly realistic VR), as a use case.