Human Factors Testing for Medical Devices: FDA Requirements, IEC 62366, and Validation Testing Guide (2026)
The complete guide to human factors testing for medical devices — FDA guidance, IEC 62366-1 usability engineering, formative and summative evaluations, sample size requirements, HF submission categories, and how to prepare documentation that passes regulatory review.
Why Human Factors Testing Is Now Mandatory for Medical Devices
A nurse programs an infusion pump to deliver 10 mL/hour of heparin, but the interface makes it easy to accidentally enter 100 mL/hour. A patient uses an autoinjector at home but fails to hold it against the skin long enough for the full dose to be delivered. A clinician misreads a pulse oximeter display because the numbers are too small and the alarm is indistinguishable from ambient hospital noise.
These are not hypothetical scenarios — they represent the types of use errors that cause real patient harm every day. Human factors engineering (HFE) — also called usability engineering — exists to prevent these errors by designing medical devices that people can use safely, effectively, and intuitively.
Regulators worldwide now treat human factors as a patient safety imperative, not a nice-to-have. The FDA requires human factors validation testing data in premarket submissions for many devices. EU Notified Bodies expect a Usability Engineering File (UEF) under IEC 62366-1 as part of MDR conformity assessment. China's NMPA has issued its own medical device usability engineering guidance. And the FDA's FY2026 guidance agenda includes final guidance on "Content of Human Factors Information in Medical Device Marketing Submissions," signaling even more specific documentation expectations.
This guide explains the complete human factors testing framework for medical devices — what regulators require, how to plan and conduct testing, what to document, and how to prepare submissions that pass regulatory review.
Human Factors Engineering vs. Usability Engineering
The terms "human factors engineering" and "usability engineering" are often used interchangeably, but there is a subtle distinction:
- Human factors engineering (HFE): The broader discipline of understanding how humans interact with systems, considering physical, cognitive, and organizational factors. The FDA uses this term.
- Usability engineering (UE): A subset of human factors that focuses specifically on the user interface — how effectively, efficiently, and satisfactorily users can achieve their goals. IEC 62366 uses this term.
In practice, for medical device regulation, both terms refer to the same core process: systematically analyzing, designing, and validating the user interface of a medical device to minimize use-related risks.
The Regulatory Framework
FDA Requirements
The FDA's approach to human factors is built on three pillars:
1. Design Controls (21 CFR 820.30 / QMSR)
Human factors is embedded in the design control requirements. Under the FDA's Quality Management System Regulation (QMSR, effective February 2, 2026), which incorporates ISO 13485:2016, design controls require:
- Design input: User needs and intended use must be defined
- Design verification: Confirmation that design outputs meet design inputs
- Design validation: Confirmation that the device conforms to defined user needs and intended uses — this is where human factors validation testing fits
2. FDA Guidance: "Applying Human Factors and Usability Engineering to Medical Devices" (2016)
This is the FDA's primary guidance on human factors. It describes:
- When human factors validation testing is expected
- How to conduct formative evaluations during development
- How to design and execute summative (validation) testing
- What to include in a human factors engineering report for FDA submissions
The FDA also has a separate guidance specifically for combination products: "Application of Human Factors Engineering Principles for Combination Products."
3. Draft Guidance: "Content of Human Factors Information in Medical Device Marketing Submissions"
This draft guidance (a B-List final guidance topic for FY2026) provides specific instructions on what human factors documentation to include in different types of marketing submissions. It introduces a risk-based categorization system with three HF submission categories, each requiring different levels of documentation.
IEC 62366-1:2015 + Amendment 1:2020
IEC 62366-1 is the international standard for usability engineering of medical devices. It is recognized by the FDA as a consensus standard, meaning a declaration of conformity to IEC 62366-1 can satisfy part of the FDA's premarket review requirements.
Key elements of IEC 62366-1:
- Usability engineering process: A structured process for identifying use-related hazards, designing mitigations, and validating the user interface
- Risk management alignment: Direct links to ISO 14971 risk management
- Formative evaluation: Iterative evaluations during development to identify and address usability issues
- Summative evaluation: Final evaluation to demonstrate that the user interface is safe for intended users in the intended use environment
IEC 62366-2:2016 is a technical report that provides guidance on implementing IEC 62366-1, including detailed descriptions of usability engineering methods.
ANSI/AAMI HE75:2025
This standard provides comprehensive guidance on human factors engineering for medical device design, covering user characteristics, use environment considerations, and testing methods. It complements IEC 62366-1 and is recognized by the FDA.
EU MDR Requirements
Under the EU Medical Device Regulation, Notified Bodies expect manufacturers to:
- Maintain a Usability Engineering File (UEF) as part of the technical documentation
- Demonstrate that the usability engineering process has been followed per IEC 62366-1
- Show traceability from hazards to design controls to validation evidence
- Address usability in the risk management file per ISO 14971
The Human Factors Engineering Process
Step 1: Identify Intended Users, Uses, and Use Environments
Define:
- User profiles: Who will use the device? (Surgeons, nurses, patients, caregivers, technicians)
- User characteristics: Training level, physical capabilities, cognitive abilities, sensory limitations
- Use environments: Operating rooms, ICUs, patient homes, ambulances, clinics
- Use conditions: Lighting, noise, interruptions, time pressure, personal protective equipment
Step 2: Conduct Task Analysis
Break down every user interaction with the device into discrete tasks. For each task:
- Identify critical tasks — those where a use error could result in serious harm
- Identify frequent tasks — those performed regularly
- Identify difficult tasks — those requiring complex cognitive or physical actions
Critical tasks are the focus of summative (validation) testing because they carry the highest risk.
Step 3: Identify Use-Related Hazards and Risks
Using the task analysis and user profiles, identify:
- Use errors: Actions (or inactions) that differ from what is expected
- Close calls: Near-miss situations
- Use difficulties: Situations that are confusing or error-prone but may not lead to harm
- Positive surprises: Unexpectedly easy or intuitive aspects
Connect each use-related hazard to the risk management file per ISO 14971.
Step 4: Design Mitigations
Implement design changes to reduce use-related risks. The hierarchy of controls applies:
- Design to eliminate: Remove the hazard entirely (e.g., physically prevent incompatible connections)
- Design to reduce: Make the error less likely (e.g., confirmatory prompts, distinct alarm sounds)
- Protective measures: Guards, interlocks, or automatic shutoffs
- Information for safety: Labeling, instructions for use, training
The FDA expects manufacturers to rely primarily on design-based controls rather than labeling or training alone.
Formative Evaluation: Testing During Development
What Is Formative Evaluation?
Formative evaluations are iterative assessments conducted during the design process to identify usability issues and inform design improvements. They are not pass/fail — their purpose is to learn and improve.
Methods
| Method | When to Use | Participants | Typical Duration |
|---|---|---|---|
| Cognitive walkthrough | Early design, paper prototypes | 2–5 HF experts | 1–2 days |
| Heuristic evaluation | Early-to-mid design | 3–5 usability experts | 1–3 days |
| Formative usability test | Mid-to-late design, working prototypes | 5–8 representative users per group | 1–2 weeks |
| Comparative evaluation | When choosing between design alternatives | 5–8 representative users | 1–2 weeks |
| Expert review | Any stage | 1–3 HF experts | 1–3 days |
Best Practices for Formative Evaluation
- Conduct multiple rounds: Plan for at least 2–3 formative evaluations during development
- Test with real users: Include representative end users, not just colleagues or internal staff
- Simulate realistic conditions: Test in environments that approximate the actual use environment
- Document findings and design changes: Maintain a traceability log linking findings to design modifications
Summative (Validation) Testing: Demonstrating Safety
What Is Summative Evaluation?
Summative evaluation — called human factors validation testing by the FDA — is the final assessment conducted after design is complete to demonstrate that the device can be used safely and effectively by intended users in realistic conditions.
This is the test that generates the data submitted to the FDA as part of a 510(k), De Novo, or PMA submission.
Sample Size Requirements
| Regulatory Authority | Minimum Participants per User Group |
|---|---|
| FDA | 15 participants per distinct user group |
| IEC 62366-1 | Representative and sufficient (no specific number mandated, but testing is the expected method) |
| NMPA (China) | 20 participants per user group |
What constitutes a user group: Distinct user populations who perform different tasks or have different characteristics. For example:
- A hospital infusion pump might have two user groups: nurses (15 participants) and biomedical engineers (15 participants) = 30 total
- A home-use autoinjector might have three user groups: adult patients, elderly patients, and caregivers = 45 total
The FDA recommends discussing the planned user groups and sample sizes during a Pre-Submission (Q-Submission) meeting before conducting summative testing.
Test Design
A well-designed summative evaluation includes:
1. Participant Recruitment
- Recruit participants who represent the intended user population
- Include users across the range of experience levels (novice to experienced)
- Account for physical and cognitive limitations relevant to the use environment (e.g., test with gloves on in simulated OR conditions)
2. Test Environment
- Simulate the actual use environment as closely as possible
- Include realistic distractions (alarms, interruptions, noise)
- Use actual or simulated device versions that represent the final design
3. Test Protocol
- Define critical tasks that will be evaluated
- Establish objective performance metrics (task completion, errors, time)
- Predefine acceptance criteria for each critical task
- Include both performance tasks (hands-on device use) and knowledge tasks (comprehension questions)
4. Data Collection
- Objective data: Task completion rates, error rates, error types, time on task
- Subjective data: User satisfaction, perceived difficulty, confidence ratings
- Observational data: Detailed observations of use errors, close calls, and use difficulties
- Root cause analysis: For every use error, determine why it occurred
5. Root Cause Analysis
For every observed use error during summative testing, conduct a root cause analysis:
- Was the error due to the device design? (User interface, controls, displays, alarms)
- Was it due to the use environment? (Lighting, noise, interruptions)
- Was it due to user characteristics? (Training, experience, physical limitations)
- Was it due to the instructions for use?
Acceptance Criteria
The summative evaluation does not require zero errors — no device is perfectly usable. Instead, demonstrate that:
- Critical task errors are understood: Every use error on a critical task has been identified, root-caused, and assessed for risk
- Residual risk is acceptable: The residual risk of use-related hazards is acceptable per the manufacturer's risk management policy and ISO 14971
- Design is defensible: The manufacturer can explain why the current design represents the best practical balance of safety, effectiveness, and usability
What to Document: The Human Factors Engineering Report
FDA Submission Content
The FDA expects a Human Factors Engineering Report (also called a Summary Report) in the marketing submission. Based on the FDA's draft guidance on "Content of Human Factors Information in Medical Device Marketing Submissions," the report should include:
HF Submission Category 1: Devices Without Critical Tasks
For devices where no use-related hazards could lead to serious harm, provide:
- A conclusion and high-level summary stating that no critical tasks were identified
- Rationale for the determination
HF Submission Category 2: Devices With Critical Tasks, No New Validation Testing Needed
For modified devices where existing validation data remains applicable, provide:
- Conclusion and high-level summary
- Description of the device modifications
- Rationale for why existing human factors data remains valid
- Summary of known use-related problems
HF Submission Category 3: Devices Requiring New Validation Testing
For new devices or modified devices where critical tasks are affected, provide the complete report:
- Conclusion and high-level summary: Overall assessment of device usability
- Device description: Intended users, uses, and use environments
- Summary of use-related risk analysis: Critical tasks, identified hazards, and risk control measures
- Summary of formative evaluations: Methods, participants, key findings, and design changes made
- Summative evaluation details: Full protocol, participant demographics, test environment description
- Results: Task performance data, error rates, root cause analysis
- Residual risk assessment: Analysis of remaining use-related risks and justification for acceptability
- Conclusions: Overall determination that the device is safe and effective for intended users
Usability Engineering File (UEF) — EU MDR
For EU submissions, maintain a Usability Engineering File that includes:
- User profiles and use environment descriptions
- Context of use analysis
- Task analysis (identifying critical tasks)
- Hazard and risk analysis linking to ISO 14971
- Formative evaluation findings and design decisions
- Summative evaluation protocol and results
- Traceability matrix: hazards → controls → evidence
- Residual risk rationale
The UEF must be available for Notified Body review during conformity assessment.
When Is Human Factors Testing Required?
FDA Expectations
The FDA expects human factors validation testing for:
| Device Type | HF Testing Expected? |
|---|---|
| Devices with critical tasks (where use errors could cause serious harm) | Yes — summative testing required |
| Combination products (autoinjectors, inhalers, prefilled syringes) | Yes — mandatory per combination product guidance |
| Home-use devices (patients with limited training) | Yes — strongly expected |
| Surgical devices (used in high-stress environments) | Yes — strongly expected |
| Software with clinical decision support | Yes — for AI-enabled devices specifically |
| Simple, low-risk devices (e.g., tongue depressors) | No — not expected |
| Modified devices where critical tasks are unchanged | May not require new testing (HF Submission Category 2) |
Specific FDA Guidance Documents
The FDA has issued multiple guidance documents relevant to human factors:
- "Applying Human Factors and Usability Engineering to Medical Devices" (2016) — The primary HFE guidance
- "Application of Human Factors Engineering Principles for Combination Products" — Specific to drug-device and biologic-device combination products
- "Comparative Analyses and Related Comparative Use Human Factors Studies for a Drug-Device Combination Product Submitted in an ANDA" — For generic combination products
- "Content of Human Factors Information in Medical Device Marketing Submissions" (draft, B-List for FY2026) — Risk-based documentation categories
- "Artificial Intelligence-Enabled Device Software Functions" (draft, January 2025) — HF expectations for AI-enabled devices
AI-Enabled Devices: Special Human Factors Considerations (2026)
For AI-enabled medical devices, the FDA's January 2025 draft guidance introduces additional human factors requirements:
- Cognitive load assessment: Evaluate whether the AI's outputs add to or reduce cognitive burden on clinicians
- Transparency features: Test whether users understand what the AI is doing, what its limitations are, and how to interpret confidence levels
- Override documentation: For clinical decision support, evaluate whether clinicians can effectively override AI recommendations and whether the override process is documented
- Model card comprehension: Test whether users can understand and act on the information in the AI model card
These requirements align with the broader 2026 regulatory trend toward explainable AI (XAI) and model transparency in medical devices.
Common Reasons Human Factors Submissions Fail
Based on FDA feedback and industry experience:
No summative testing conducted: The most common deficiency — manufacturers submit formative data only, without the summative validation study that demonstrates safety
Missing critical task identification: Failure to identify all critical tasks means some use-related hazards are not evaluated
Inadequate user representation: Testing with convenience samples (colleagues, engineers) rather than representative end users
Unrealistic test conditions: Testing in quiet, well-lit offices rather than simulated clinical environments with distractions
Missing root cause analysis: Observing errors during testing but not analyzing why they occurred or what design changes could prevent them
No discussion of residual risk: The FDA expects manufacturers to acknowledge and justify any residual use-related risks, not claim the device is "error-free"
Insufficient sample size: Fewer than 15 participants per user group for FDA submissions
Practical Implementation Checklist
| Phase | Action | Key Deliverable |
|---|---|---|
| Planning | Define user profiles, use environments, and critical tasks | User Research Plan |
| Analysis | Conduct task analysis and use-related risk analysis | Use-Related Risk Analysis Report |
| Design | Implement design mitigations for identified hazards | Design Controls Traceability |
| Formative 1 | Initial usability evaluation with prototypes | Formative Evaluation Report 1 |
| Iteration | Redesign based on formative findings | Updated Design Documentation |
| Formative 2 | Second evaluation of improved design | Formative Evaluation Report 2 |
| Final design | Lock design for validation | Final Design Baseline |
| Summative | Conduct human factors validation testing | Summative Evaluation Report |
| Documentation | Prepare HFE Report for submission | Complete HFE Submission Package |
| Pre-submission | Discuss HF approach with FDA (recommended) | FDA Q-Submission Meeting Package |
Frequently Asked Questions
Is human factors testing required for 510(k) submissions?
It depends on the device. The FDA expects human factors validation testing for devices with critical tasks where use errors could cause serious harm — including combination products, home-use devices, and surgical devices. For simple, low-risk devices without critical tasks, a statement explaining why HF testing is not needed may suffice.
How many participants do I need for summative testing?
The FDA recommends a minimum of 15 participants per distinct user group. If your device has two user groups (e.g., nurses and patients), you need at least 30 participants total. China's NMPA recommends 20 per user group.
Can I use the same summative study for both FDA and EU?
Yes, provided the study meets the requirements of both frameworks. IEC 62366-1 does not specify a minimum sample size, but if you test with 15 participants per user group (meeting the FDA requirement), this will also satisfy EU Notified Body expectations.
What is the difference between formative and summative evaluation?
Formative evaluations are iterative, conducted during development, and used to improve the design. They are not pass/fail. Summative evaluations (validation testing) are conducted on the final design to demonstrate that the device can be used safely. Summative testing requires a formal protocol, predefined acceptance criteria, and comprehensive documentation.
Do I need to conduct human factors testing for software-only devices (SaMD)?
Yes, if the software has a user interface and critical tasks where use errors could lead to serious harm. For AI-enabled SaMD, the FDA's January 2025 draft guidance adds specific requirements for cognitive load assessment, transparency testing, and override documentation.
What happens if users make errors during summative testing?
Errors during summative testing are expected and do not automatically mean the device fails. The key is to identify all errors, analyze their root causes, assess the residual risk, and justify why the risk is acceptable. If errors reveal design deficiencies that create unacceptable risk, the manufacturer should modify the design and conduct additional testing.
Key Takeaways
- Human factors testing is now a regulatory requirement, not a best practice — the FDA, EU, and other regulators expect evidence that devices can be used safely by intended users in realistic conditions
- The FDA recommends 15 participants per user group for summative (validation) testing; China's NMPA recommends 20
- IEC 62366-1 provides the international standard for usability engineering and is recognized by the FDA as a consensus standard
- The FDA's draft guidance on "Content of Human Factors Information in Marketing Submissions" introduces three risk-based categories that determine what documentation to include
- AI-enabled medical devices have additional human factors requirements including cognitive load assessment, transparency testing, and model card comprehension
- Successful human factors submissions require a complete process: user research → task analysis → risk identification → formative evaluations → design iteration → summative validation → comprehensive documentation
- The most common reason submissions fail is the absence of summative validation testing — formative evaluations alone are not sufficient