FDA 510(k) Substantial Equivalence Decision Tool
Use this utility to pressure-test whether your substantial equivalence (SE) argument is defensible before you lock your submission package. The goal is practical: identify weak points early, map each difference to objective evidence, and avoid preventable review-cycle delays.
This page is a planning aid, not legal advice. Teams should validate final strategy against current FDA requirements and product-specific risk context.
Build a Stronger 510(k) Submission Narrative
Cruxi helps teams structure substantial equivalence logic, align evidence to claims, and deliver coherent eSTAR-ready 510(k) submissions.
Explore 510(k) Submission ServicesTool 1: SE Argument Strength Calculator
Scorecard Inputs (0 to 5 each)
Use conservative scoring. A "3" should mean your team can defend the point with traceable evidence today, not after future cleanup.
Run the scorecard to view your weighted SE strength score, risk band, and highest-priority remediation targets.
Tool 2: Difference-to-Evidence Mapping Matrix
A frequent SE failure pattern is describing differences without clear evidence linkage. Use this matrix format to make reviewer logic explicit.
| Difference Category | Observed Difference | Potential Safety/Effectiveness Question | Evidence to Close Gap | Owner + Due Date |
|---|---|---|---|---|
| Indications / use context | Subject use setting differs from predicate wording. | Could use context alter risk profile or clinical relevance? | Line-by-line indications comparison + rationale memo. | RA lead, pre-lock milestone |
| Core technology | Modified energy delivery parameters. | Do changes introduce new hazard modes? | Bench performance protocol, acceptance criteria, summary table. | R&D + V&V owner |
| Materials / patient contact | New material in patient-contacting path. | Any new biocompatibility concern? | Biocompatibility rationale and test evidence mapped to contact profile. | Toxicology / biocomp owner |
| Software / workflow | New automation feature in user workflow. | Could user behavior or failure modes shift materially? | Hazard analysis update, validation summary, usability rationale. | Software + HF lead |
| Labeling / IFU | Warnings differ from predicate emphasis. | Any mismatch between risk controls and user instructions? | Cross-reference table between risks, mitigations, and IFU warnings. | RA + labeling owner |
What Reviewers Need to See Clearly
1) Claims-to-Evidence Traceability
Every major claim in your SE narrative should resolve to a named test, data package, or rationale source. If a reviewer cannot trace a claim quickly, expect follow-up questions.
2) Difference Handling Logic
Listing differences is not enough. Explain why each difference does not create a new safety/effectiveness question and identify the data that supports that conclusion.
3) Internal Consistency
Inconsistency across indications, device description, IFU, and testing claims is a common avoidable risk driver in review cycles.
SE Narrative Readiness Checklist
- task_altPrimary predicate rationale approved by regulatory, technical, and quality stakeholders.
- task_altEach technology difference linked to explicit risk rationale and evidence package.
- task_altIndications language stays consistent across forms, narrative, and labeling drafts.
- task_altOpen gaps have assigned owners, due dates, and acceptance criteria.
- task_altFinal narrative reviewed against current FDA pathway expectations before file lock.
Related 510(k) Resources
- link510(k) Submission Services for end-to-end filing support.
- linkPredicate Analysis Scorecard for predicate-fit and risk screening.
- linkRTA Deficiency Risk Estimator to prioritize pre-file quality work.
- linkLabeling and IFU Guide for consistency checks across your submission.
- linkFDA Database Research Guide for predicate and regulatory data workflows.
Frequently Asked Questions
What makes an SE narrative weak in FDA review?
Most weak narratives fail on logic continuity: they mention differences but do not close them with clear evidence. Low traceability and inconsistent claims create avoidable review friction.
Do we need clinical data for substantial equivalence?
Not always. Teams often use non-clinical data when differences are bounded and evidence is sufficient. Clinical data may become necessary if non-clinical evidence cannot answer key safety or effectiveness questions.
When should teams reconsider the 510(k) strategy?
If your highest-impact differences still generate unresolved safety/effectiveness questions after planned testing, escalate early and reassess pathway strategy before final submission lock.