In general, it's just not economical to test every possible use case. In addition, there's often insufficient business case to justify a fix for some defects (eg. suitable work-arounds exist or the negative consequences are insignificant compared to the cost of fixing). All software (and therefore most hardware) ships with bugs, it's just a question of the number, severity, and the use cases which trigger them that determines the perceived quality.
point taken. but how hard is it to run a comparative benchmark from this version to the version it is super-seeding?
the coders/devs should have built it in such a way that the software can be improved each time, not the reverse!
i only hope that the bad benchmarks are because some sort of new programming techniques (or anything) has been used and it needs more work to fully bring out the true quality. /dream