Exactly how do you think Nvidia is raking in $Billions on H100 sales, if AI is so heavily CPU bottle-necked? Wouldn't it be Intel who would be swamped with demand and a 1.5 year manufacturing backlog?
You know that deep learning has quite a lot of the smartest wet brains working on it. It's been a buzzword for probably 6 or 7 years, now. Do you think it didn't occur to anyone that using Python might be slow?
Finally, consider that PyTorch absorbed Caffe2, which was the successor to the Caffe framework - both of which are C++ -based. Do you think PyTorch would've achieved such dominance and succeeded in eclipsing Caffe and subsuming the C++ interface of Caffe2, if it were as slow as you seem to think? Not only that, but there are frameworks not based on Python - how did they not dominate the Python-based ones?
These are just some sanity-checks that you should be doing, before you jump to such sweeping conclusions. You don't even need to delve down to the level of graph compilation, LLVM, or runtime code generation to suspect that all may not be as it seems.
When you search for something that seems to confirm your existing conclusion, it's called
confirmation bias, which can lead to self-deception.
Instead, what you should do is search for answers to the original question. In fact, by searching for confirmation that Python is slow, you've missed two fundamental points, as well as some tertiary details:
- No, deep learning doesn't predominantly rely on Python performance.
- Even code which does rely on Python performance isn't necessarily interpreted. There are multiple ways you can compile Python to run as native code (i.e. not just bundling an interpreter, but actually compiling it like C++).
- Python performance is a moving target. By looking at data from 2017, you're missing the continual optimizations that have been made to the official Python interpreter. Also, performance aggregates are just that. If you have a more specific workload in mind, then you should look at benchmarks for that workload.
What I find funny is that your source isn't even fundamentally addressing the issue of Python performance or from a reputable source with expertise on such matters. What's worse is that it cites a paper published 6 years ago. You don't seem to weight the quality of your source nearly as much as the fact that it agrees with your preconceptions. Did you even check the cited paper to look at its methodology or whether the article might be taking its findings out of context?