1. No, it's performed using the Tensor cores, not the shaders.
2. Try what? What does decompression have to do with GPU performance? CPU decompression is still way faster than GPU decompression. So who cares.
3. Fine. Not sure where the RT question came into play, because you and I were talking about Ampere's FP compute being largely useless in gaming workloads.
4. Of course it's real. If you can fill all the ALUs you can get lot of compute performance. That 'certain path' just isn't games. Most people here don't care about compute workloads. And even if you did, you'd probably prefer getting RTX A6000 instead of a 3090 because it would have pro drivers that fully extract Amperes compute capabilities.
5. Cool. So what?
Nvidia will use their tensor cores for ML, because they're faster at matrix and bfloat operations than cuda cores or regular vector ALUs.
6. Modified how? Do you have the white papers for RDNA2 and Series X's ISA to be able to confirm that DX12U features enabled on XSX doesn't exist in PC RDNA2? I want to see the receipts. You can't make a spurious claim here and have no evidence.
7. This sentence is unintelligible. What are you talking about? TIOPS 'hide' extra performance? What is being 'hidden' and how? This sentence makes absolutely no sense whatsoever.
It's still 40% slower than Navi 21 so who cares how many TeraFLOPS it has. Not sure what relevance Turing and it's compute performance has on a conversation about Ampere and Navi 21.