QA/Test
Manage and conduct quality assurance and testing activities (e.g. requirements testing, code review, regression testing, user acceptance).
Improvement Planning
Practices-Outcomes-Metrics (POM)
Representative POMs are described for QA/Test at each level of maturity.
- 2Basic
- Practice
- Develop test scenarios for IT solutions post-development, automating in some instances.
- Outcome
- Test cases are developed after the IT solutions are implemented, and data generated may undergo limited review.
- Metrics
- % of solutions with test scenarios developed post release.
- % of automated unit and acceptance test scenarios.
- Practice
- Address quality issues for high-stake solutions.
- Outcome
- Quality may not be consistent for all solutions.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Use informal user acceptance testing to get input and feedback on test and validation from some stakeholders.
- Outcome
- Many needs are met by informal user acceptance testing.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Implement some automation of regression/integration tests and a common set of tools to support testing, such as debugging tools.
- Outcomes
- Manual tests are replaced by automation and a common set of tools.
- Tests are recorded and repeatable.
- Metrics
- % of test effort allocated to manual testing.
- % of manual testing that is exploratory.
- 3Intermediate
- Practice
- Define unit and acceptance tests before solutions are constructed, with input and review from stakeholders.
- Outcomes
- Efficient and effective unit and acceptance test processes are available after solutions are constructed, with input and review from stakeholders.
- Tests may include non-functional aspects such as performance, compliance, and security, for both in-house and third-party components.
- Metrics
- % of automated unit and acceptance test scenarios.
- % of stakeholders involved in defining unit and acceptance tests.
- Test execution speed (capacity to meet CI/CD requirements).
- % of high-risk requirements covered by automated testing.
- Practice
- Measure code quality using static code analysis tools.
- Outcomes
- The quality of designs are good.
- Fault slip-through levels are low.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Define and agree acceptance criteria with users.
- Outcome
- Rich user feedback is captured — regarding work as done only if it passes acceptance tests.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practices
- Ensure that tests are routinely automated, including many non-functional tests.
- Record and trace results.
- Outcomes
- Automated test suites run efficiently to provide results rapidly.
- Test are traceable back to their requirement.
- Metrics
- % of test effort allocated to manual testing.
- % of manual testing that is exploratory.
- 4Advanced
- Practice
- Routinely review and update tests to maintain comprehensive coverage and include editorial aspects such as plain language and graphics criteria in test scope.
- Outcomes
- Efficient and effective test processes including editorial and graphical considerations.
- Well-defined test data management strategies enable realistic scenario testing.
- Metrics
- % of automated test scenarios including editorial and graphical considerations.
- % of tests reviewed for improvement per annum.
- Test execution speed (capacity to meet CI/CD requirements).
- % of high-risk requirements covered by automated testing.
- Practice
- Ensure that test code undergoes review and evaluation.
- Outcome
- Test code is robust and extensible.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Continue to refine test processes to reflect user input.
- Outcome
- Tests are more robust — reflecting rich user feedback.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practices
- Ensure all tests are routinely automated, and provide rapid results that are traceable.
- Extend automation to all aspects of the solution including, for example, readability analysis.
- Outcomes
- Pervasive automated test suites consistently provide results rapidly.
- Test are fully traceable back to their requirement.
- Metrics
- % of test effort allocated to manual testing.
- % of manual testing that is exploratory.
- 5Optimized
- Practice
- Continually review and optimize tests using innovative methods, when available.
- Outcome
- Continuously refreshed representative test scenarios using anonymized real-life test data improve automation and test coverage.
- Metrics
- % of automated test scenarios including editorial and graphical considerations.
- % of tests reviewed for improvement per annum.
- Test execution speed (capacity to meet CI/CD requirements).
- % of high-risk requirements covered by automated testing.
- Practice
- Ensure frequent graduated releases.
- Outcome
- More frequent smaller releases allow for deeper feedback from active use and shorten the feedback loop, producing a quicker response and excellent quality of solutions.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Carry out post-deployment observation and interaction sessions to ensure processes and solutions are working effectively.
- Outcome
- Test processes are working seamlessly to ensure a high-quality solution.
- Metrics
- # of design-related incidents.
- User acceptance rate (%).
- Practice
- Experiment with innovative methods and solutions to improve the level of automation and efficiency of tests on an ongoing basis.
- Outcome
- There is optimal use of automation at all times.
- Metrics
- % of test effort allocated to manual testing.
- % of manual testing that is exploratory.