We have to consider the fact that A100 is designed as a compute card for specialised applications such as AI. Its not exactly a gpu, not a quadro.
I'm not familiar with the benchmark, but I got the impression it merely uses CUDA for computation. In that sense, the specs should be entirely predictive of the resultant performance.
The Titan V has 653 GB/sec of memory bandwidth and 14.9 TFLOPS of fp32 compute (at boost clocks).
The A100 has 1555 GB/sec of memory bandwidth and 19.5 TFLOPS of fp32 compute (at ??? clocks).
So, if the workload is compute-limited, then it should be at least 30.9% faster (possibly more, depending on whether the A100 figures I saw were base or boost). However, if it's limited by memory bandwidth, then it should be 138% faster! Either way, 11.2% is really disappointing. I think it likely points to bottlenecks in the software, but we don't really know. It's definitely not good
news - that's for sure!