During my career as a product designer, I struggled with a recurring question: What parts of a design truly need testing — and what can safely rely on design patterns and best practices?
Sometimes I'd lean heavily on conventions and proven patterns. Other times, I'd feel unsure and test everything "just to be safe." And then came the sign-off meetings.
A stakeholder would ask: "Did we test this flow?" or "Are we confident users understand this interaction?"
And honestly — I didn't always know how to answer.
That uncertainty pushed me to rethink how I approach testing.
The Real Problem: Testing Everything Isn't Practical
In theory, everything should be tested. In reality, time, budget, and access to users are limited.
Over-testing slows teams down. Under-testing creates risk.
What I needed wasn't more testing — but a clear way to decide what deserves it.
A Simple Framework That Changed My Approach

Over time, I developed a simple but effective framework to decide what to test and how much confidence I can have without testing.
I call it the Must / Should / Maybe Test Framework.
This framework helps me:
- Prioritize testing effort
- Explain decisions clearly to stakeholders
- Defend design choices with confidence
1. Must Test: High Risk, High Impact
These are the areas I always push to test.
Must test if the design:
- Introduces a new or unfamiliar interaction
- Is critical to conversion, revenue, or retention
- Has serious consequences if misunderstood
Examples:
- Checkout or payment flows
- Onboarding that blocks user access
- Destructive actions (delete, cancel, unsubscribe)
Here, best practices are not enough. Real user feedback is non-negotiable.
2. Should Test: Medium Risk, Some Uncertainty
These are areas where patterns exist — but context matters.
Should test if the design:
- Combines multiple known patterns in a new way
- Targets a new user segment
- Has multiple possible interpretations
Examples:
- Filtering and sorting logic
- Empty states with strong calls-to-action
- Complex forms
If time allows, I test. If not, I validate through reviews, heuristics, or quick internal checks — and I'm transparent about it.
3. Maybe Test: Low Risk, High Confidence
These are areas where testing adds little value.
Maybe test if the design:
- Uses well-established design patterns
- Has low impact if misunderstood
- Is purely cosmetic
Examples:
- Button placement following conventions
- Standard input fields
- Minor visual adjustments
- Icon choices with labels
Here, best practices + design systems + experience are usually enough.
How This Helped in Stakeholder Conversations

Now, when someone asks: "Did we test this?" — I don't panic.
I can say: "This fell into our 'Should test' category. We didn't run user testing due to time constraints, but we validated it using proven patterns and cross-functional reviews."
Or: "This is a 'Must test' flow, and here's what we learned from users."
The conversation shifts from defensiveness to decision-making.
Why This Framework Works
This approach:
- Creates shared expectations
- Builds stakeholder trust
- Reduces over-testing
- Makes design decisions more intentional
It turns testing from a checkbox into a strategic tool.
Good design isn't about testing everything. It's about knowing what needs evidence, what needs judgment, and what needs speed.
Having a clear testing criteria helped me move from uncertainty to confidence — and made sign-off meetings a lot less stressful.