FDA AI-Enabled Device Predicate Mining Method: How to Identify, Evaluate, and Defend Your Predicate for 510(k) and De Novo
A methodical approach to mining FDA's AI-enabled device list and 510(k) database for predicate devices — covering technological characteristic extraction, public summary limitations, weak predicate argument avoidance, and a complete predicate evaluation matrix.
What This Article Covers / Does Not Cover
This article covers one specific task: how to use FDA's public AI-enabled medical device list, the 510(k) database, and publicly available summary documents to identify, evaluate, and defend a predicate device for an AI-enabled medical device seeking 510(k) clearance or De Novo classification. It includes a predicate mining workflow, a technological characteristic extraction table, a predicate strength scoring matrix, and common failure modes specific to AI-enabled devices.
This article does not cover the general 510(k) submission process, how to write a substantial equivalence argument from scratch, or how to prepare performance testing data. For the general 510(k) framework, see 510(k) Submission Guide. For predicate device selection fundamentals, see 510(k) Predicate Device Guide. For the broader AI/ML regulatory landscape, see AI/ML Medical Device Regulatory Guide.
Why AI-Enabled Device Predicate Mining Is Different
Approximately 97% of FDA-authorized AI-enabled medical devices have been cleared through the 510(k) pathway. Through the end of 2025, FDA's list contains over 1,450 cumulative authorized AI-enabled devices. In 2025 alone, 295 AI/ML-enabled devices received clearance, with 211 (71.5%) in radiology, 26 (8.8%) in cardiovascular, and 14 (4.7%) in neurology.
But AI-enabled devices present unique predicate challenges:
- Software opacity: Public 510(k) summary statements rarely describe AI architecture, model type, training data characteristics, or inference pipeline details.
- Rapid evolution: A predicate cleared two years ago may use fundamentally different technology than what is standard today.
- Mixed technology comparisons: FDA permits AI-enabled devices to be found substantially equivalent to non-AI-enabled predicates if no new questions of safety and effectiveness are raised (per FDA's draft AI-enabled device software functions guidance, January 2025).
- Sparse non-radiology predicates: Outside radiology, the predicate pool shrinks dramatically, forcing sponsors toward De Novo or toward weaker predicate arguments.
Step 1: Define Your Technological Characteristic Profile
Before searching for predicates, document your own device's technological characteristics in a structured format. FDA evaluates substantial equivalence along two axes: intended use and technological characteristics. The table below provides the characteristic categories most relevant to AI-enabled devices.
Technological Characteristic Extraction Table
| Characteristic Category | Data to Document | Source in Your Documentation |
|---|---|---|
| Intended use / indications | Specific clinical indication, patient population, clinical setting, user type | IFU, intended use statement |
| AI function type | Detection, triage, quantification, segmentation, diagnosis, treatment recommendation, workflow optimization | Software design specification |
| Input data modality | CT, MRI, X-ray, ultrasound, ECG, pathology slide, clinical text, physiological waveform | Software requirements spec |
| Anatomy / physiology | Organ system, anatomical region, physiological parameter | Device description |
| AI/ML approach | Deep learning (CNN, transformer, etc.), classical ML, rule-based, hybrid | Algorithm design document |
| Output type | Binary classification, multi-class, bounding box, segmentation mask, continuous score, text report | Output specification |
| Integration | Standalone SaMD, embedded in hardware, PACS-integrated, EHR-integrated, cloud-deployed, edge-deployed | Architecture document |
| Clinical workflow position | Triaging/screening, diagnostic aid, clinical decision support, second read, quality check | Intended use / clinical evaluation |
| Human oversight | Human-in-the-loop, human-over-the-loop, autonomous | Risk management file |
| Training data characteristics | Dataset size, demographics, multi-site vs single-site, ground truth method | AI/ML performance validation report |
| Performance metrics | Sensitivity, specificity, AUROC, Dice coefficient, PPV, NPV, F1 — with confidence intervals | Clinical/performance validation |
| PCCP inclusion | Whether a predetermined change control plan is included | PCCP document |
| Product code | FDA product code(s) applicable | Classification determination |
| Review panel | Radiology, Cardiovascular, Neurology, etc. | Classification database |
Step 2: Mine the FDA AI-Enabled Device List
2.1 Access the List
FDA publishes the "Artificial Intelligence-Enabled Medical Devices" list on its website at fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-enabled-medical-devices. The list is downloadable in CSV, Excel, and XML formats. It contains:
- Date of final decision
- Submission number (e.g., K252379)
- Device name
- Company name
- Panel (lead review division)
- Primary product code
2.2 Filter Strategically
Apply filters in this order:
| Filter Priority | Field | Why |
|---|---|---|
| 1 | Primary product code | Narrows to your device's regulatory family |
| 2 | Panel (review division) | Ensures same clinical domain |
| 3 | Date of final decision | Prefer predicates cleared within the last 5 years |
| 4 | Company name | Same-manufacturer predicates simplify comparison |
| 5 | Device name keyword | Functional similarity screening |
2.3 Record Candidate Predicates
For each candidate identified, create a predicate candidate card:
| Field | Your Device | Candidate Predicate |
|---|---|---|
| 510(k) / De Novo number | (to be assigned) | K25XXXXX |
| Device name | [Your device] | [Predicate device] |
| Company | [Your company] | [Predicate company] |
| Product code | [Your code(s)] | [Predicate code] |
| Panel | [Your panel] | [Predicate panel] |
| Intended use | [Yours] | [From public summary] |
| Clearance date | — | [Date] |
| Still marketed? | — | [To verify] |
Step 3: Extract Technological Characteristics from Public Summaries
3.1 What Public Summaries Actually Contain
FDA's publicly available 510(k) summary statements (found in the 510(k) database) typically include:
- Device description (sometimes detailed, sometimes vague)
- Intended use statement
- Comparison to predicate(s): intended use, technological characteristics
- Summary of performance testing
- Brief statement on software validation
- Conclusion on substantial equivalence
3.2 What Public Summaries Almost Never Contain
| Missing Information | Impact on Your SE Argument |
|---|---|
| AI model architecture details | Cannot verify algorithmic equivalence |
| Training dataset size/composition | Cannot compare training rigor |
| Specific performance metrics with CIs | Cannot do quantitative comparison |
| Preprocessing pipeline details | May need to assume differences |
| Post-processing logic | Often a source of hidden differences |
| Cybersecurity controls | Increasingly scrutinized for AI devices |
| Cloud/edge deployment architecture | Relevant for cybersecurity and performance |
3.3 Technological Characteristic Comparison Matrix
Build this matrix for every candidate predicate. Use "Same," "Different," or "Unknown" for each row:
| Characteristic | Your Device | Predicate | Assessment | Evidence Source |
|---|---|---|---|---|
| Intended use | [Yours] | [Theirs] | Same/Different | Public summary |
| Input data modality | [Yours] | [Theirs] | Same/Different | Public summary |
| AI function type | [Yours] | [Theirs/Unknown] | Same/Different/Unknown | Public summary or inferred |
| Anatomy/physiology | [Yours] | [Theirs] | Same/Different | Public summary |
| Output type | [Yours] | [Theirs/Unknown] | Same/Different/Unknown | Public summary |
| Clinical workflow position | [Yours] | [Theirs/Unknown] | Same/Different/Unknown | Inferred from IFU |
| Human oversight model | [Yours] | [Theirs/Unknown] | Same/Different/Unknown | Often not stated |
| Software architecture | [Yours] | [Theirs/Unknown] | Different/Unknown | Almost never stated |
| Performance level | [Yours with CIs] | [Theirs/Unknown] | Comparable/Unknown | Public summary if available |
Step 4: Predicate Strength Scoring
Score each candidate predicate on a 1-5 scale across the following dimensions. A score below the threshold means the predicate is too weak to rely on as a primary predicate.
Predicate Strength Scoring Matrix
| Dimension | Weight | 5 (Strong) | 3 (Moderate) | 1 (Weak) |
|---|---|---|---|---|
| Intended use match | 30% | Identical indications, population, setting | Overlapping but narrower or broader | Different clinical indication or population |
| Technological similarity | 25% | Same modality, function, output type | Same modality, different function or output | Different modality or fundamentally different technology |
| Product code match | 15% | Same product code | Related product code, same panel | Different product code, different panel |
| Currency | 10% | Cleared within 2 years | Cleared within 5 years | Cleared more than 5 years ago |
| Regulatory history | 10% | Clean, no recalls or safety alerts | Minor issues resolved | Recalls, warning letters, or withdrawal |
| Information availability | 10% | Detailed public summary with performance data | Basic public summary | Minimal or no public information |
Weighted score threshold: A weighted score below 3.0 indicates a weak predicate. Below 2.5, consider De Novo or a different predicate strategy.
Decision Tree: Primary Predicate vs. Multiple Predicates vs. De Novo
START: Is there a candidate with weighted score >= 3.5?
├── YES → Use as primary predicate
│ └── Are there gaps in technological characteristics?
│ ├── NO → Proceed with single-predicate SE argument
│ └── YES → Can gaps be addressed with additional reference devices?
│ ├── YES → Add secondary predicate(s) for specific features
│ └── NO → Address gaps with performance testing data
└── NO → Is there a candidate with weighted score 2.5-3.5?
├── YES → Can additional performance data bridge the gap?
│ ├── YES → Strengthen SE argument with robust testing
│ └── NO → Consider Pre-Submission meeting with FDA
└── NO → Is there ANY candidate with score >= 2.0?
├── YES → De Novo may be more appropriate
└── NO → De Novo pathway is likely required
Step 5: Avoid Weak Predicate Arguments
Common Failure Modes for AI-Enabled Device Predicates
| Failure Mode | What It Looks Like | Why FDA Rejects It | How to Fix It |
|---|---|---|---|
| Functional mismatch | Your device detects lung nodules; predicate triages chest X-rays | Different intended use despite same anatomy | Find a predicate with matching AI function, or do De Novo |
| "AI as feature" argument | Claiming AI is just a different implementation of the same clinical function | If AI introduces new analysis capability, it may raise new questions of safety/effectiveness | Demonstrate that AI output is clinically equivalent to predicate's non-AI output |
| Stale predicate | Predicate cleared in 2018 with outdated technology | FDA may question whether the predicate reflects current standard of care | Use a more recent predicate or provide evidence that the older predicate is still clinically relevant |
| Unknown performance gap | Cannot determine predicate's sensitivity/specificity from public data | FDA may request additional clinical validation | Conduct clinical validation study; consider Pre-Submission |
| Cross-panel predicate | Your device is Cardiovascular; predicate is Radiology | Different review panels, different clinical expertise | Find same-panel predicate or engage FDA early via Pre-Submission |
| Multiple predicate assembly | Cobbling together 3+ predicates for different features | SE argument becomes incoherent; each predicate must share the same intended use | Use one primary predicate for intended use; reference devices for specific tech comparisons only |
| Non-AI to AI gap | Predicate is non-AI; your device adds AI analysis | FDA's January 2025 draft guidance permits this IF no new safety/effectiveness questions | Provide performance data showing AI output is equivalent or superior to predicate's manual process |
Reviewer Objection Table
| Reviewer Objection | Typical FDA Language | How to Respond |
|---|---|---|
| Different intended use | "The subject device's intended use is broader than the predicate" | Narrow your indications or find a better-matched predicate |
| New questions of safety | "The AI component introduces a new mechanism of analysis not present in the predicate" | Provide clinical validation data, risk analysis, and performance comparison |
| Insufficient comparison | "The technological characteristics comparison does not adequately address differences in the software algorithm" | Supplement with detailed algorithm description, V&V data, and performance testing |
| Performance gap | "The submitted performance data does not demonstrate comparable safety and effectiveness" | Conduct additional clinical validation; consider De Novo |
Step 6: Use Pre-Submission to De-Risk Predicate Strategy
FDA's Pre-Submission program (Q-Submission) allows you to get feedback on your predicate strategy before committing to a full 510(k) or De Novo submission. For AI-enabled devices, this is especially valuable when:
- Your best predicate scores between 2.5 and 3.5 on the strength matrix
- You are considering a non-AI to AI comparison
- You are outside radiology (where predicates are sparse)
- You plan to use multiple predicates
Pre-Submission Predicate Package Contents
| Document | Purpose |
|---|---|
| Proposed intended use statement | For FDA to confirm intended use alignment |
| Predicate identification table | Show candidate predicates with scoring rationale |
| Technological comparison matrix | Highlight similarities and differences |
| Proposed performance testing plan | Show how you plan to address differences |
| Specific questions for FDA | Numbered, focused questions about predicate acceptability |
Illustrative Pre-Submission question format:
"We have identified [K-number] as our primary predicate device. Our device shares the same intended use (detection of [condition] from [modality]), same input data type, and same clinical workflow position. The primary technological difference is [describe]. We plan to address this difference with [proposed testing]. Does FDA agree that this predicate and testing approach are appropriate for a 510(k) submission?"
Step 7: Document the Predicate Strategy in Your Submission
Evidence Traceability Table
| Submission Section | Predicate-Related Content | Evidence Records |
|---|---|---|
| Device description | Side-by-side device description comparison | Device description document, predicate public summary |
| Substantial equivalence | Intended use comparison, technological characteristics comparison, performance comparison | SE narrative, comparison tables |
| Software documentation | Software architecture comparison (where available), V&V summary | IEC 62304 documentation, software description |
| Performance testing | Bench/clinical testing bridging technological differences | Test reports, statistical analysis |
| Clinical evaluation | Clinical validation study results vs. predicate performance (where available) | Clinical evaluation report, validation study report |
| Labeling comparison | IFU comparison table | IFU/labeling documents |
RACI for Predicate Mining Process
| Task | Regulatory Affairs | Engineering/Clinical | Quality | Management |
|---|---|---|---|---|
| Define technological profile | C | R | I | A |
| Search FDA AI device list | R | C | I | I |
| Extract public summary data | R | C | I | I |
| Score candidate predicates | R | C | I | A |
| Conduct Pre-Submission (if needed) | R | C | I | A |
| Build SE comparison tables | R | C | I | I |
| Review final predicate strategy | C | C | C | A |
Common Failure Modes and Remediation
| Failure Mode | Root Cause | Remediation |
|---|---|---|
| Predicate no longer marketed | Did not verify market status before submission | Check predicate company website, FDA registration listings, and distributor catalogs before committing to a predicate |
| Predicate recalled after selection | Did not monitor predicate regulatory status during submission preparation | Set up FDA recall database alerts for selected predicates |
| Inadequate public information | Chose predicate with minimal public summary | Prefer predicates with detailed public summaries; contact predicate manufacturer for additional information if appropriate |
| AI function differs from predicate | Assumed AI analysis is interchangeable with non-AI analysis | Provide clinical validation showing equivalence; consider De Novo if the AI function is fundamentally new |
| Product code mismatch | Did not verify classification before selecting predicate | Use FDA Product Classification Database to confirm product code assignment |
| Weak multiple-predicate argument | Assembled predicates that do not share intended use | Ensure all predicates share the same intended use; use reference devices only for specific technological comparisons |
Sources
- FDA. "Artificial Intelligence-Enabled Medical Devices." Updated through December 2025. fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-enabled-medical-devices
- FDA. "Artificial Intelligence-Enabled Device Software Functions: Lifecycle Considerations and Premarket Review Recommendations." Draft Guidance, January 2025.
- Innolitics. "2025 Year in Review: AI/ML Medical Device 510(k) Clearances." innolitics.com/articles/year-in-review-ai-ml-medical-device-k-clearances/
- IntuitionLabs. "FDA's AI Medical Device List: Stats, Trends & Regulation." Updated March 2026.
- FDA. "The 510(k) Program: Evaluating Substantial Equivalence in Premarket Notifications." Guidance.
- Complizen. "What Is Substantial Equivalence (SE) in FDA 510(k)? Definition & Criteria." 2025.