MedDeviceGuideMedDeviceGuide
Back

FDA AI-Enabled Device Predicate Mining Method: How to Identify, Evaluate, and Defend Your Predicate for 510(k) and De Novo

A methodical approach to mining FDA's AI-enabled device list and 510(k) database for predicate devices — covering technological characteristic extraction, public summary limitations, weak predicate argument avoidance, and a complete predicate evaluation matrix.

Ran Chen
Ran Chen
Global MedTech Expert | 10× MedTech Global Access
2026-05-0514 min read

What This Article Covers / Does Not Cover

This article covers one specific task: how to use FDA's public AI-enabled medical device list, the 510(k) database, and publicly available summary documents to identify, evaluate, and defend a predicate device for an AI-enabled medical device seeking 510(k) clearance or De Novo classification. It includes a predicate mining workflow, a technological characteristic extraction table, a predicate strength scoring matrix, and common failure modes specific to AI-enabled devices.

This article does not cover the general 510(k) submission process, how to write a substantial equivalence argument from scratch, or how to prepare performance testing data. For the general 510(k) framework, see 510(k) Submission Guide. For predicate device selection fundamentals, see 510(k) Predicate Device Guide. For the broader AI/ML regulatory landscape, see AI/ML Medical Device Regulatory Guide.


Why AI-Enabled Device Predicate Mining Is Different

Approximately 97% of FDA-authorized AI-enabled medical devices have been cleared through the 510(k) pathway. Through the end of 2025, FDA's list contains over 1,450 cumulative authorized AI-enabled devices. In 2025 alone, 295 AI/ML-enabled devices received clearance, with 211 (71.5%) in radiology, 26 (8.8%) in cardiovascular, and 14 (4.7%) in neurology.

But AI-enabled devices present unique predicate challenges:

  1. Software opacity: Public 510(k) summary statements rarely describe AI architecture, model type, training data characteristics, or inference pipeline details.
  2. Rapid evolution: A predicate cleared two years ago may use fundamentally different technology than what is standard today.
  3. Mixed technology comparisons: FDA permits AI-enabled devices to be found substantially equivalent to non-AI-enabled predicates if no new questions of safety and effectiveness are raised (per FDA's draft AI-enabled device software functions guidance, January 2025).
  4. Sparse non-radiology predicates: Outside radiology, the predicate pool shrinks dramatically, forcing sponsors toward De Novo or toward weaker predicate arguments.

Step 1: Define Your Technological Characteristic Profile

Before searching for predicates, document your own device's technological characteristics in a structured format. FDA evaluates substantial equivalence along two axes: intended use and technological characteristics. The table below provides the characteristic categories most relevant to AI-enabled devices.

Technological Characteristic Extraction Table

Characteristic Category Data to Document Source in Your Documentation
Intended use / indications Specific clinical indication, patient population, clinical setting, user type IFU, intended use statement
AI function type Detection, triage, quantification, segmentation, diagnosis, treatment recommendation, workflow optimization Software design specification
Input data modality CT, MRI, X-ray, ultrasound, ECG, pathology slide, clinical text, physiological waveform Software requirements spec
Anatomy / physiology Organ system, anatomical region, physiological parameter Device description
AI/ML approach Deep learning (CNN, transformer, etc.), classical ML, rule-based, hybrid Algorithm design document
Output type Binary classification, multi-class, bounding box, segmentation mask, continuous score, text report Output specification
Integration Standalone SaMD, embedded in hardware, PACS-integrated, EHR-integrated, cloud-deployed, edge-deployed Architecture document
Clinical workflow position Triaging/screening, diagnostic aid, clinical decision support, second read, quality check Intended use / clinical evaluation
Human oversight Human-in-the-loop, human-over-the-loop, autonomous Risk management file
Training data characteristics Dataset size, demographics, multi-site vs single-site, ground truth method AI/ML performance validation report
Performance metrics Sensitivity, specificity, AUROC, Dice coefficient, PPV, NPV, F1 — with confidence intervals Clinical/performance validation
PCCP inclusion Whether a predetermined change control plan is included PCCP document
Product code FDA product code(s) applicable Classification determination
Review panel Radiology, Cardiovascular, Neurology, etc. Classification database

Recommended Reading
FDA eSTAR Electronic Submission Template: Complete Guide to 510(k) and De Novo Submissions
510(k) Regulatory2026-04-17 · 16 min read

Step 2: Mine the FDA AI-Enabled Device List

2.1 Access the List

FDA publishes the "Artificial Intelligence-Enabled Medical Devices" list on its website at fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-enabled-medical-devices. The list is downloadable in CSV, Excel, and XML formats. It contains:

  • Date of final decision
  • Submission number (e.g., K252379)
  • Device name
  • Company name
  • Panel (lead review division)
  • Primary product code

2.2 Filter Strategically

Apply filters in this order:

Filter Priority Field Why
1 Primary product code Narrows to your device's regulatory family
2 Panel (review division) Ensures same clinical domain
3 Date of final decision Prefer predicates cleared within the last 5 years
4 Company name Same-manufacturer predicates simplify comparison
5 Device name keyword Functional similarity screening

2.3 Record Candidate Predicates

For each candidate identified, create a predicate candidate card:

Field Your Device Candidate Predicate
510(k) / De Novo number (to be assigned) K25XXXXX
Device name [Your device] [Predicate device]
Company [Your company] [Predicate company]
Product code [Your code(s)] [Predicate code]
Panel [Your panel] [Predicate panel]
Intended use [Yours] [From public summary]
Clearance date [Date]
Still marketed? [To verify]

Step 3: Extract Technological Characteristics from Public Summaries

3.1 What Public Summaries Actually Contain

FDA's publicly available 510(k) summary statements (found in the 510(k) database) typically include:

  • Device description (sometimes detailed, sometimes vague)
  • Intended use statement
  • Comparison to predicate(s): intended use, technological characteristics
  • Summary of performance testing
  • Brief statement on software validation
  • Conclusion on substantial equivalence

3.2 What Public Summaries Almost Never Contain

Missing Information Impact on Your SE Argument
AI model architecture details Cannot verify algorithmic equivalence
Training dataset size/composition Cannot compare training rigor
Specific performance metrics with CIs Cannot do quantitative comparison
Preprocessing pipeline details May need to assume differences
Post-processing logic Often a source of hidden differences
Cybersecurity controls Increasingly scrutinized for AI devices
Cloud/edge deployment architecture Relevant for cybersecurity and performance

3.3 Technological Characteristic Comparison Matrix

Build this matrix for every candidate predicate. Use "Same," "Different," or "Unknown" for each row:

Characteristic Your Device Predicate Assessment Evidence Source
Intended use [Yours] [Theirs] Same/Different Public summary
Input data modality [Yours] [Theirs] Same/Different Public summary
AI function type [Yours] [Theirs/Unknown] Same/Different/Unknown Public summary or inferred
Anatomy/physiology [Yours] [Theirs] Same/Different Public summary
Output type [Yours] [Theirs/Unknown] Same/Different/Unknown Public summary
Clinical workflow position [Yours] [Theirs/Unknown] Same/Different/Unknown Inferred from IFU
Human oversight model [Yours] [Theirs/Unknown] Same/Different/Unknown Often not stated
Software architecture [Yours] [Theirs/Unknown] Different/Unknown Almost never stated
Performance level [Yours with CIs] [Theirs/Unknown] Comparable/Unknown Public summary if available

Step 4: Predicate Strength Scoring

Score each candidate predicate on a 1-5 scale across the following dimensions. A score below the threshold means the predicate is too weak to rely on as a primary predicate.

Predicate Strength Scoring Matrix

Dimension Weight 5 (Strong) 3 (Moderate) 1 (Weak)
Intended use match 30% Identical indications, population, setting Overlapping but narrower or broader Different clinical indication or population
Technological similarity 25% Same modality, function, output type Same modality, different function or output Different modality or fundamentally different technology
Product code match 15% Same product code Related product code, same panel Different product code, different panel
Currency 10% Cleared within 2 years Cleared within 5 years Cleared more than 5 years ago
Regulatory history 10% Clean, no recalls or safety alerts Minor issues resolved Recalls, warning letters, or withdrawal
Information availability 10% Detailed public summary with performance data Basic public summary Minimal or no public information

Weighted score threshold: A weighted score below 3.0 indicates a weak predicate. Below 2.5, consider De Novo or a different predicate strategy.

Decision Tree: Primary Predicate vs. Multiple Predicates vs. De Novo

START: Is there a candidate with weighted score >= 3.5?
├── YES → Use as primary predicate
│   └── Are there gaps in technological characteristics?
│       ├── NO → Proceed with single-predicate SE argument
│       └── YES → Can gaps be addressed with additional reference devices?
│           ├── YES → Add secondary predicate(s) for specific features
│           └── NO → Address gaps with performance testing data
└── NO → Is there a candidate with weighted score 2.5-3.5?
    ├── YES → Can additional performance data bridge the gap?
    │   ├── YES → Strengthen SE argument with robust testing
    │   └── NO → Consider Pre-Submission meeting with FDA
    └── NO → Is there ANY candidate with score >= 2.0?
        ├── YES → De Novo may be more appropriate
        └── NO → De Novo pathway is likely required

Recommended Reading
EU AI Act + MDR Single Evidence Matrix: How to Build One Combined Technical File Without Duplicating Work
EU MDR / IVDR Digital Health & AI2026-05-05 · 17 min read

Step 5: Avoid Weak Predicate Arguments

Common Failure Modes for AI-Enabled Device Predicates

Failure Mode What It Looks Like Why FDA Rejects It How to Fix It
Functional mismatch Your device detects lung nodules; predicate triages chest X-rays Different intended use despite same anatomy Find a predicate with matching AI function, or do De Novo
"AI as feature" argument Claiming AI is just a different implementation of the same clinical function If AI introduces new analysis capability, it may raise new questions of safety/effectiveness Demonstrate that AI output is clinically equivalent to predicate's non-AI output
Stale predicate Predicate cleared in 2018 with outdated technology FDA may question whether the predicate reflects current standard of care Use a more recent predicate or provide evidence that the older predicate is still clinically relevant
Unknown performance gap Cannot determine predicate's sensitivity/specificity from public data FDA may request additional clinical validation Conduct clinical validation study; consider Pre-Submission
Cross-panel predicate Your device is Cardiovascular; predicate is Radiology Different review panels, different clinical expertise Find same-panel predicate or engage FDA early via Pre-Submission
Multiple predicate assembly Cobbling together 3+ predicates for different features SE argument becomes incoherent; each predicate must share the same intended use Use one primary predicate for intended use; reference devices for specific tech comparisons only
Non-AI to AI gap Predicate is non-AI; your device adds AI analysis FDA's January 2025 draft guidance permits this IF no new safety/effectiveness questions Provide performance data showing AI output is equivalent or superior to predicate's manual process

Reviewer Objection Table

Reviewer Objection Typical FDA Language How to Respond
Different intended use "The subject device's intended use is broader than the predicate" Narrow your indications or find a better-matched predicate
New questions of safety "The AI component introduces a new mechanism of analysis not present in the predicate" Provide clinical validation data, risk analysis, and performance comparison
Insufficient comparison "The technological characteristics comparison does not adequately address differences in the software algorithm" Supplement with detailed algorithm description, V&V data, and performance testing
Performance gap "The submitted performance data does not demonstrate comparable safety and effectiveness" Conduct additional clinical validation; consider De Novo

Step 6: Use Pre-Submission to De-Risk Predicate Strategy

FDA's Pre-Submission program (Q-Submission) allows you to get feedback on your predicate strategy before committing to a full 510(k) or De Novo submission. For AI-enabled devices, this is especially valuable when:

  1. Your best predicate scores between 2.5 and 3.5 on the strength matrix
  2. You are considering a non-AI to AI comparison
  3. You are outside radiology (where predicates are sparse)
  4. You plan to use multiple predicates

Pre-Submission Predicate Package Contents

Document Purpose
Proposed intended use statement For FDA to confirm intended use alignment
Predicate identification table Show candidate predicates with scoring rationale
Technological comparison matrix Highlight similarities and differences
Proposed performance testing plan Show how you plan to address differences
Specific questions for FDA Numbered, focused questions about predicate acceptability

Illustrative Pre-Submission question format:

"We have identified [K-number] as our primary predicate device. Our device shares the same intended use (detection of [condition] from [modality]), same input data type, and same clinical workflow position. The primary technological difference is [describe]. We plan to address this difference with [proposed testing]. Does FDA agree that this predicate and testing approach are appropriate for a 510(k) submission?"


Step 7: Document the Predicate Strategy in Your Submission

Evidence Traceability Table

Submission Section Predicate-Related Content Evidence Records
Device description Side-by-side device description comparison Device description document, predicate public summary
Substantial equivalence Intended use comparison, technological characteristics comparison, performance comparison SE narrative, comparison tables
Software documentation Software architecture comparison (where available), V&V summary IEC 62304 documentation, software description
Performance testing Bench/clinical testing bridging technological differences Test reports, statistical analysis
Clinical evaluation Clinical validation study results vs. predicate performance (where available) Clinical evaluation report, validation study report
Labeling comparison IFU comparison table IFU/labeling documents

RACI for Predicate Mining Process

Task Regulatory Affairs Engineering/Clinical Quality Management
Define technological profile C R I A
Search FDA AI device list R C I I
Extract public summary data R C I I
Score candidate predicates R C I A
Conduct Pre-Submission (if needed) R C I A
Build SE comparison tables R C I I
Review final predicate strategy C C C A

Recommended Reading
FDA ASCA Test Report Acceptance Package: How to Build a Bulletproof Evidence Package for 510(k) and De Novo Submissions
510(k) Regulatory2026-05-05 · 22 min read

Common Failure Modes and Remediation

Failure Mode Root Cause Remediation
Predicate no longer marketed Did not verify market status before submission Check predicate company website, FDA registration listings, and distributor catalogs before committing to a predicate
Predicate recalled after selection Did not monitor predicate regulatory status during submission preparation Set up FDA recall database alerts for selected predicates
Inadequate public information Chose predicate with minimal public summary Prefer predicates with detailed public summaries; contact predicate manufacturer for additional information if appropriate
AI function differs from predicate Assumed AI analysis is interchangeable with non-AI analysis Provide clinical validation showing equivalence; consider De Novo if the AI function is fundamentally new
Product code mismatch Did not verify classification before selecting predicate Use FDA Product Classification Database to confirm product code assignment
Weak multiple-predicate argument Assembled predicates that do not share intended use Ensure all predicates share the same intended use; use reference devices only for specific technological comparisons

Sources

  • FDA. "Artificial Intelligence-Enabled Medical Devices." Updated through December 2025. fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-enabled-medical-devices
  • FDA. "Artificial Intelligence-Enabled Device Software Functions: Lifecycle Considerations and Premarket Review Recommendations." Draft Guidance, January 2025.
  • Innolitics. "2025 Year in Review: AI/ML Medical Device 510(k) Clearances." innolitics.com/articles/year-in-review-ai-ml-medical-device-k-clearances/
  • IntuitionLabs. "FDA's AI Medical Device List: Stats, Trends & Regulation." Updated March 2026.
  • FDA. "The 510(k) Program: Evaluating Substantial Equivalence in Premarket Notifications." Guidance.
  • Complizen. "What Is Substantial Equivalence (SE) in FDA 510(k)? Definition & Criteria." 2025.