Quality Assurance, Confidence Levels, and Testing
I agree on the 'quality assurance' comment -- semantically and practically it is not a very useful term.
(a) semantically, any product will have an inherent level of Quality [0 <= Q <= 1] as soon as it is released in whichever environment. A level of Quality is already 'assured' as soon as the software product exits the compiler.
(b) practically, a 100% assurance of High Quality [a score of 1.00] is impossible, not even on a quantum level -- as soon as we observe something at that level, it changes the specifics of what we were trying to observe.
However, on the confidence/acceptance scale, it must be agreeable that Testing directly/indirectly ascertains the Confidence level perceived from using the SUT(Subject Under Test). This would be attested to by 'tested' cars, planes, and rockets -- and by extension, software products. These products are not perfect 100% Confidence Level ["C" score = 1.00] but empirically also not Confidence Level 0. Their C Level is somewhere between 0 and 1 ; the lower the score, the more likely the product is not up to expectations. And the converse is also true.
In conclusion, it would seem every usable product out there has reached a certain level of confidence/acceptance -- and that would be directly or indirectly through the efforts of Testing.
In most cases, Confidence/Acceptance Level can be perceived as an inference of Quality. The two appear to be correlated but semantically not the same. Confidence/Acceptance is UX-centric, while Quality is SUT-centric.

Comments
Post a Comment