43:34 are threads acessing memory with constant speed? because if they go out of sync then there is additional performance penalty I guess you even admit that 5 minutes later without mentioning that it will affect total bandwidth as well I guarantee that if you implement same read pattern with same thread timing using only one socket you get terrible bandwidth as well because the limiting factor is not inter-socket interaction but DDR structure itself
What is precisely meant by no interaction between NUMA nodes if only talking to L3 cache at ruclips.net/video/f0ZKBusa4CI/видео.htmlfeature=shared&t=2005, in terms of effect on memory bandwidth?
Great presentation. Would love to see the rest!
Awesome talk! Lots of intricate details and valuable insights. Thank you!
Very glad to hear that you enjoyed it! Thank you for your comments.
Great talk, shame can't see the rest
NUMA seemed like a reasonable solution to continue providing higher memory throughput. Assuming the software is aware of the NUMA.
I'd like to add NUMA awareness to the EM solve openEMS which is severely memory-bandwidth
bound
Great talk sad it got cutout
43:34 are threads acessing memory with constant speed? because if they go out of sync then there is additional performance penalty I guess
you even admit that 5 minutes later without mentioning that it will affect total bandwidth as well
I guarantee that if you implement same read pattern with same thread timing using only one socket you get terrible bandwidth as well because the limiting factor is not inter-socket interaction but DDR structure itself
What is precisely meant by no interaction between NUMA nodes if only talking to L3 cache at ruclips.net/video/f0ZKBusa4CI/видео.htmlfeature=shared&t=2005, in terms of effect on memory bandwidth?
36:25 "the ratio here is about 50"
how in the hell it could be fifty? it's a fraction of the cell height which is 10x
the Y is logarithmic
24:20 this graph would definitely look better in totals, not per thread
Funny numa in Japanese means “swamp” : D