They formed a business model on untested software that wasn't intended for mainstream parts in the first place. So now they are crapping their pants there are bugs on software that was just released (making it new) when the stable proven software stack has been released for years.
Let me give you an example of why this is stupid.
Ideally you want a vector comparison that returns the closest value that returns 1 in similarity. Well you might have thousands of vectors to test to find the closest ideal one. You might have a software stack item that supports comparing 256 vectors at a time (internally) because DEDICATED AI hardware can perform a bitroll on 256 vectors and resolve which one is greatest in log2(256) = 8 clock ticks. But consumer hardware might only be able to compare 16 vectors at a time. This creates tournament rounds of multiple passes which takes more clock ticks. Well the internal stack might not take this into consideration and if the next uOp operates on the previous result based on the fact it will take 8 clock ticks, you get an incorrect answer as the mulitple round tournament isn't complete.
When you have quite literally hundreds or even over 1000 uOps, comparing the timing operations of all instructions becomes a very NP Complete problem and it takes time to solve as the domain space for testing exponentially grows based on the number of permutations for instruction orders increases. Each instruction can break down into dozens of uOps and just because one instruction breaks down to steps x1 y1 z1, the exact same instruction next time will break down to x2, y3, z20 (even though uOps are the same) because the resources are busy for x1 y1, z1. Then you have to verify they complete in the order expected (Data dependency race condition) Some uOps can have a variable amount of execution time and it's up to the OOE scheduler to account for this. NOW GP GPUs don't have OOE Schedulars. But their software stack can emulate them.