Flagship whitepaper · Methodology · ~16 min read
Critical Testing Processes (CTP) is the framework we use to assess and improve a client's test function. Unlike TMMi, CMMI-DEV, or IEEE 29119, CTP does not impose a staged maturity model. It names twelve testing processes that determine whether a test team succeeds or fails, defines what each process should accomplish, and leaves the order of improvement to the organization's own pain points and business priorities.
This paper introduces the twelve processes, the four criteria we used to select them, how CTP assessments work in practice, and how the model adapts to continuous-delivery organizations today. The checklists and companion articles under the QA Library are the operational artifacts of this framework.
The premise
When we wrote the book Critical Testing Processes in the early 2000s, we started from a simple premise: in any test team, some processes are critical and some are not. The critical ones are the ones that determine whether the team succeeds or fails; the non-critical ones are tractable with ordinary engineering effort once the criticals are in place. A framework for test-process improvement should name the critical ones, describe what each should accomplish, and leave everything else alone.
That premise put CTP at odds with the large-enterprise process-improvement models of the day — TMMi (Test Maturity Model integration) and TPI Next, and by extension CMMI-DEV. Those models all share a common structural feature: they define staged maturity levels (typically five), each with prerequisites that must be satisfied before an organization can "advance" to the next level. A process area's importance to your organization is secondary to where it sits in the maturity staircase.
What's wrong with staged maturity? In practice, businesses want to make improvements based on two things: the business value of the improvement, and the organizational pain the improvement will alleviate. A staged model may direct you to invest in a low-value, low-pain process area simply because it appears at your current maturity level. That's a misallocation of scarce improvement capacity. Worse, it creates an incentive to optimize for the rating rather than the underlying outcome — a pattern sometimes called "maturity model theater."
CTP is non-prescriptive by design. It tells you what a good process does, not when in a journey you should build it, and it measures your actual processes against business needs rather than a maturity checklist.
What makes a testing process "critical"
We applied four criteria to the universe of things a test team does. A process is critical if it meets all four:
- Frequency. The process is repeated often enough that quality differences compound into large differences in team efficiency.
- Cooperation. The process involves many people, often cross-functionally, so quality differences affect team cohesion and organizational trust.
- Visibility. The process is visible to peers and superiors, so quality differences affect test-team credibility.
- Link to project success. The process has a direct line to project outcomes — either test-team or project-team effectiveness.
Put differently: a critical testing process directly and significantly affects the test team's ability to find bugs, build confidence, reduce risk, and generate information the business can act on. Teams that do these processes well usually succeed; teams that do them poorly usually fail regardless of individual talent.
Applying these criteria to the full testing mandate produced twelve critical testing processes.
The twelve processes
1. Testing (the macro process)
The overall testing activity viewed at the strategic level. This is not an additional process on top of the other eleven; it's the umbrella that ties them together and connects them to the business. A good overall testing process balances analytical (risk-based, requirements-based, checklist-based) and dynamic (exploratory, attack-based, error-guessing) strategies, aligns with the organization's delivery lifecycle, and produces measurable value. Everything below is a constituent of this macro process.
2. Establishing context
Aligns testing with the project and the organization. Clarifies expectations across stakeholders, tailors every other testing process to the specific engagement, and captures the test policy and test strategy when those artifacts earn their keep.
We do not prescribe that a test policy document must exist — in some organizations it's essential, in others it's bureaucratic overhead. The test-policy-and-strategy question is a policy choice, not a deliverable mandate. The Context Discovery Process checklist is the operational artifact here.
3. Quality risk analysis
Identifies the risks to system quality that testing will address. Aligns testing with those risks. Builds stakeholder consensus on what is to be tested (and how much) and what is not (and why). QRA is the input that makes every downstream testing activity either relevant or wasteful.
Operational artifacts: the Quality Risk Analysis Process checklist, the QRA whitepaper, and the FMEA template. The whitepaper on Risk Perception and Cognitive Bias covers the facilitation discipline needed to run QRA sessions that produce a risk register worth trusting.
4. Test estimation
Balances the cost and duration of testing against project needs and risks. Produces an accurate, actionable, and flexible forecast of testing work. Demonstrates the return on the testing investment in terms the business recognizes. Good estimation is a political as much as a technical activity; a test manager who cannot defend an estimate will not get the resources needed to execute.
Operational artifacts: the Test Estimation Process checklist and the Test Estimation article, which together cover WBS, divide-and-conquer, Delphic Oracle, Three-Point, and Wideband Delphi techniques plus the modern asymmetries introduced by AI-assisted development and sprint-cadence delivery.
5. Test planning
Builds consensus and commitment across the test team and the broader project team. Creates a map detailed enough to coordinate test participants and thin enough to remain usable under schedule pressure. Captures context for retrospectives and future projects.
Operational artifacts: the Test Plan Template.
6. Test team development
Testing is as good as the team that does it. This process matches test-team skills to the critical test tasks, assures competence in the skill areas the project actually requires, and continuously aligns team capability with the organization's evolving use of testing. Today that means explicit tracks for API and data testing, cloud and infrastructure testing, AI/ML testing and evaluation, security testing, and observability-driven quality work, on top of classical functional and domain test skills.
Operational artifacts: the Team Building Process checklist and job descriptions for Test Manager, Automation Test Engineer, and Test Role Exercises.
7. Test system development
Ensures coverage of the critical quality risks. Creates tests that reproduce the customer's and user's experiences of quality. Balances resource and time requirements against the criticality of the risk. Covers test cases, test data, test procedures, test environments, and all supporting material.
Operational artifacts: the Test Case Templates, the Thoughts on Test Data whitepaper, and the Managing Complex Test Environments companion.
8. Test release management
If the test object isn't in the test environment, you can't test it. If each release to test isn't better than the one before, you're on the wrong trajectory. This process focuses on how to get solid, reliable releases into the test environment — which today typically means CI/CD pipelines delivering containerized builds to ephemeral environments, rather than weekly full-environment installs from a release-management org.
Operational artifacts: the Test Release Process checklist and the Release Management Processes article.
9. Test execution
The running of test cases and comparison of results against expectations. Generates the information about bugs, working behavior, and broken behavior that is the fundamental value of testing. Consumes significant resources. In most lifecycles, sits on or near the critical path to release.
Operational artifacts: the Test Execution Process checklist and the Test Execution Processes article.
10. Bug reporting
Turns the output of execution into an opportunity to improve the system. Delivers part of the value of testing to the project team — the individual contributors and line managers — and builds tester credibility with engineering.
Operational artifacts: the Bug Reporting Process ten-step checklist and the Bug Reporting Processes deep dive.
11. Results reporting
Provides leadership with the information needed to guide the project. Delivers another part of the value of testing — to line managers, senior managers, and executives. Separates the message from the messenger when test results are bad news. Builds tester credibility with management.
Operational artifacts: the Test Results Reporting Process checklist, the Effective Test Status Reporting article, and the Risk-Based Test Results Reporting article which covers residual-risk trend charts as the executive artifact.
12. Change management
Allows the test team and the project team to respond to what they've learned. Selects the right changes in the right order. Focuses effort on the highest-ROI activities. In CTP, change management is part of the test function's job — not a separate PMO function — because change decisions have large immediate consequences for the test plan.
Operational artifacts: the Change Management Process checklist.
How CTP assessments work
CTP-based test-process improvement begins with an assessment of the existing test process against the twelve processes. That assessment produces a set of prioritized improvement recommendations based on organizational needs, not a maturity model.
A CTP assessment is always tailored to the engagement. We have performed:
- Narrow focused assessments — a single test team, sometimes a single process area (e.g. only "test system development"), producing a detailed improvement plan for that scope.
- Team-wide assessments — the full test function across the organization, producing a multi-quarter improvement roadmap.
- Enterprise-wide assessments — all processes that affect software quality, including upstream requirements and design processes, producing a cross-functional improvement program.
Quantitative metrics we evaluate
Assessments weigh multiple quantitative signals, selected from the organization's own instrumentation:
- Defect Detection Effectiveness (DDE) — percentage of defects caught in a given test phase versus defects that leak past it. 85% industry baseline, 95% achievable. See the Metrics Part 2 whitepaper for the full methodology including phase containment across the lifecycle.
- Return on the testing investment — cost of testing vs. value delivered through prevented field failures, support deflection, and accelerated feature delivery. See the Investing in Software Testing series for the canonical ROI model.
- Requirements coverage and risk coverage — the two complementary measures of whether testing aimed at the right things.
- Test release overhead — time and cost spent per release delivered to test. A proxy for the maturity of the delivery pipeline.
- Defect report rejection rate — the percentage of tester-submitted reports closed without a fix (invalid, duplicate, won't-fix-at-priority). A balanced indicator: very high rates suggest poor bug reports or low engineering trust; very low rates sometimes signal a prioritization bar that's too permissive.
Modern assessments layer in the DORA four (deployment frequency, lead time, change failure rate, mean time to restore), flakiness and cycle time for automated suites, block rate at CI gates, and production-telemetry-derived defect escape rate. Those are modern additions; the original five remain the backbone.
Qualitative factors we evaluate
Metrics are necessary but never sufficient. Good assessments also evaluate:
- Test-team role and effectiveness — how the organization positions testing in the SDLC and whether it treats the team as a resource or a rubber stamp.
- Usefulness of the test plan — do people read the plan; does it shape behavior; does it survive contact with project reality.
- Test-team skills — in testing itself, in the business domain, and in the relevant technology stack.
- Value of defect reports — would engineering pay for them if they had to.
- Usefulness of results reports — do executives act on them; do they build or erode trust over time.
- Change-management utility and balance — is the change process selecting the right changes in the right order, or is it a bureaucracy that slows work without filtering it.
Assessment output
A CTP assessment of the kind we run for consulting clients produces a 50 to 100-page report covering each of the twelve processes, a quantitative and qualitative evaluation, specific improvement recommendations, a business justification for each recommendation, and an implementation roadmap. Two CTP assessment reports for different organizations may recommend similar changes but put them in very different order, because each client has a different opportunity and constraint profile. That's the point of a non-prescriptive model.
Figure below — the summary layout we use to present a process-by-process evaluation — makes the non-prescriptive property explicit. No process is rated "fulfilled / unfulfilled" or "maturity level N." Instead each gets a qualitative evaluation plus a momentum indicator (improving / stable / degrading), and each comes with specific improvement candidates ranked by opportunity.
CTP in a continuous-delivery organization
The original CTP work assumed a lifecycle where testing occupied a distinct phase. Today, many teams run continuous delivery: builds flow through automated quality gates into production several times a day, with no fixed "test phase" boundary. CTP adapts, with some mapping:
- "Test release management" becomes pipeline health. The critical process is whether the CI/CD pipeline reliably produces testable builds at the expected cadence, with ephemeral environments provisioned on demand. The metric is pipeline success rate and lead time, not release-note accuracy.
- "Test execution" becomes gate health and exploratory cadence. Automated tests run in the pipeline; human testers spend their time on exploratory, risk-based, and edge-case work between deploys. The metric is block rate at each gate plus time-to-signal for escaped defects.
- "Results reporting" becomes a live dashboard plus a weekly narrative. Telemetry-driven quality dashboards (Grafana/Datadog/Looker) replace static weekly reports; the narrative layer — residual quality risk trend, big-risk sentinel events, upcoming known-risk releases — remains a human artifact.
- "Change management" becomes feature-flag hygiene and release-train governance. The change being managed is frequently a flag flip or a progressive rollout increment, not a major release.
The twelve critical processes are the same. The instrumentation and cadence are different.
When CTP is the wrong choice
CTP is not the universal answer. Organizations that should consider alternatives include:
- Safety-critical regulated environments (aviation DO-178C, medical IEC 62304, automotive ISO 26262) where the applicable standards require a specific process framework and a specific artifact set. CTP principles still apply to the engineering work, but the assessment framework needs to match the certification regime.
- Organizations that are mid-CMMI-appraisal or mid-TMMi-assessment and have already committed budget and political capital to that path. Switching frameworks mid-flight costs more than it saves.
- Organizations seeking a badge for procurement or RFP reasons. CTP does not produce a maturity-level rating that can be printed on a slide. That's a feature, not a bug — but a real constraint if the rating is the point.
For everyone else — the overwhelming majority of modern software organizations — a non-prescriptive, business-driven framework that produces actionable improvements in the order the organization needs them is the better fit.
A closing note on reassessment
Process assessment is not a one-time activity. Whether you use CTP or any other framework, do the assessment, implement the highest-value improvements, and then reassess at regular intervals — a year is a reasonable cadence in most organizations, shorter when improvements are in-flight. Reassessment tells you whether the improvement actually moved the metrics, where the momentum is, and where the next increment of improvement has the highest payoff.
Test-process improvement that doesn't produce measurable, durable change in DDE, defect escape rate, cycle time, or customer-perceived quality has not improved anything. Assessment is only the first half of the work. Measurement over time is the other half.
Related resources
- QA Library — the operational checklists and templates for all twelve processes.
- Investing in Software Testing, Part 1 — the ROI model the "test estimation" and "results reporting" processes feed.
- Quality Risk Analysis whitepaper — deep dive on the QRA process.
- Metrics for Software Testing, Part 1 — the metrics framework CTP assessments use.
- Effective Test Status Reporting — companion to the "results reporting" critical process.