Skip to main content
Glama
uat-test-plan-template.md8.52 kB
# User Acceptance Testing (UAT) Plan Template ## UAT Plan: [Project/Feature Name] **Prepared By**: [Business Analyst Name] **Date Created**: [YYYY-MM-DD] **Last Updated**: [YYYY-MM-DD] **Version**: [1.0] --- ## 1. Introduction ### 1.1 Purpose [State the purpose of this UAT plan and what will be tested] ### 1.2 Scope **In Scope**: * [Feature/Module 1 to be tested] * [Feature/Module 2 to be tested] * [Feature/Module 3 to be tested] **Out of Scope**: * [What will NOT be tested in this UAT] * [Features deferred to future testing] ### 1.3 Objectives * [Objective 1: e.g., Verify the system meets business requirements] * [Objective 2: e.g., Ensure the system is ready for production deployment] * [Objective 3: e.g., Validate workflows match real-world business processes] --- ## 2. Test Environment ### 2.1 Environment Details * **Environment URL**: [e.g., https://uat.example.com] * **Environment Type**: [UAT/Staging] * **Database**: [UAT database name/instance] * **Data**: [Production-like test data / Anonymized data] ### 2.2 Access Requirements * **User Accounts**: [How testers will receive credentials] * **VPN/Network Access**: [Any special network requirements] * **Permissions**: [Role-based access for different test scenarios] ### 2.3 Test Data * **Data Preparation**: [Who prepares test data and when] * **Data Refresh**: [How often data is refreshed] * **Data Constraints**: [Limitations or special considerations] --- ## 3. Roles and Responsibilities | Role | Name | Responsibilities | | :--- | :--- | :--- | | **UAT Lead** | [Name] | Overall coordination, reporting, decision-making | | **Business SMEs** | [Names] | Execute test cases, validate business logic | | **Business Analyst** | [Name] | Prepare test cases, facilitate testing, document issues | | **Test Coordinator** | [Name] | Schedule sessions, track progress, manage defects | | **Technical Support** | [Name] | Resolve environment issues, answer technical questions | | **Product Owner** | [Name] | Final sign-off, prioritize issues | --- ## 4. Entry and Exit Criteria ### 4.1 Entry Criteria [Conditions that must be met before UAT can begin] * ☐ All development complete for in-scope features * ☐ System testing (QA) completed with no critical defects * ☐ UAT environment ready and stable * ☐ Test data loaded and verified * ☐ Test cases reviewed and approved * ☐ UAT participants trained and available ### 4.2 Exit Criteria [Conditions that must be met to complete UAT successfully] * ☐ All planned test cases executed * ☐ No open critical (P1) or high (P2) defects * ☐ All medium (P3) defects reviewed and accepted or fixed * ☐ 95% of test cases passed * ☐ Business stakeholders provide sign-off * ☐ UAT summary report completed --- ## 5. Test Scenarios and Cases ### 5.1 Test Scenario Summary | Scenario ID | Scenario Name | Business Priority | # Test Cases | Assigned To | | :--- | :--- | :--- | :--- | :--- | | TS-001 | [Scenario name] | High/Med/Low | [Count] | [Name] | | TS-002 | [Scenario name] | High/Med/Low | [Count] | [Name] | | TS-003 | [Scenario name] | High/Med/Low | [Count] | [Name] | --- ### Test Scenario: TS-001 [Scenario Name] **Business Process**: [Which business process this tests] **User Role**: [Who performs this scenario] **Priority**: High / Medium / Low **Preconditions**: * [Precondition 1] * [Precondition 2] --- #### Test Case: TC-001 [Test Case Name] **Scenario**: TS-001 **Priority**: High / Medium / Low **Test Data**: [Reference to specific test data needed] **Test Steps**: 1. [Step 1 action] * **Expected Result**: [What should happen] 2. [Step 2 action] * **Expected Result**: [What should happen] 3. [Step 3 action] * **Expected Result**: [What should happen] **Acceptance Criteria**: * [Criterion 1] * [Criterion 2] **Actual Result**: [To be filled during testing] **Status**: [Pass / Fail / Blocked / Not Tested] **Executed By**: [Name] **Date Executed**: [YYYY-MM-DD] **Defect ID**: [If failed, reference defect] --- #### Test Case: TC-002 [Test Case Name] [Repeat structure for each test case] --- ## 6. Defect Management ### 6.1 Defect Severity Definitions | Severity | Definition | Example | Response Time | | :--- | :--- | :--- | :--- | | **P1 - Critical** | System down or major feature unusable | Login fails for all users | Immediate | | **P2 - High** | Major feature impaired, workaround exists | Report generates incorrect totals | 24 hours | | **P3 - Medium** | Minor feature issue or cosmetic problem | Alignment issue in UI | 72 hours | | **P4 - Low** | Suggestions, enhancements | Improved wording request | As resources permit | ### 6.2 Defect Logging Process 1. Tester identifies issue during testing 2. Tester logs defect in [defect tracking tool] 3. Business Analyst reviews and categorizes 4. Development team triages and assigns priority 5. Fix is implemented and deployed to UAT 6. Tester retests and closes defect ### 6.3 Defect Report Template **Defect ID**: [AUTO-GENERATED] **Test Case ID**: [TC-XXX] **Summary**: [Brief description] **Severity**: [P1/P2/P3/P4] **Status**: [Open/In Progress/Fixed/Retest/Closed/Deferred] **Steps to Reproduce**: 1. [Step 1] 2. [Step 2] 3. [Step 3] **Expected Result**: [What should happen] **Actual Result**: [What actually happened] **Screenshot/Attachment**: [Link or attach] **Environment**: [Browser, OS, environment URL] **Reported By**: [Name] **Date Reported**: [YYYY-MM-DD] --- ## 7. Test Schedule | Phase | Activities | Start Date | End Date | Owner | | :--- | :--- | :--- | :--- | :--- | | **Preparation** | Test case creation, data setup | [Date] | [Date] | [Name] | | **Training** | UAT tester orientation | [Date] | [Date] | [Name] | | **Testing Round 1** | Initial test execution | [Date] | [Date] | [Name] | | **Defect Fixing** | Resolve P1/P2 defects | [Date] | [Date] | [Name] | | **Regression Testing** | Retest fixes | [Date] | [Date] | [Name] | | **Testing Round 2** | Final validation | [Date] | [Date] | [Name] | | **Sign-off** | Approval and closure | [Date] | [Date] | [Name] | **Daily Testing Sessions**: * **Time**: [e.g., 2:00 PM - 4:00 PM] * **Duration**: [e.g., 2 hours per day] * **Location**: [Virtual/In-person, meeting link] --- ## 8. Communication and Reporting ### 8.1 Status Meetings * **Frequency**: [Daily during testing / As needed] * **Attendees**: [UAT team, project manager, developers] * **Purpose**: Progress updates, issue escalation ### 8.2 Daily Status Report **Date**: [YYYY-MM-DD] **Test Cases Executed Today**: [X] **Test Cases Passed**: [X] **Test Cases Failed**: [X] **Test Cases Blocked**: [X] **New Defects Logged**: [X] **Defects Fixed and Retested**: [X] **Issues/Risks**: [Any blockers or concerns] ### 8.3 Final UAT Report To be provided at the end of UAT with: * Executive summary * Test execution metrics * Defect summary by severity * Outstanding issues and recommendations * Sign-off signatures --- ## 9. Risks and Mitigation | Risk | Probability | Impact | Mitigation Strategy | | :--- | :--- | :--- | :--- | | [Risk 1: e.g., Testers not available] | H/M/L | H/M/L | [Have backup testers identified] | | [Risk 2: e.g., Environment instability] | H/M/L | H/M/L | [Have technical support on standby] | | [Risk 3: e.g., Incomplete test data] | H/M/L | H/M/L | [Data prep checklist and early setup] | --- ## 10. Assumptions and Dependencies **Assumptions**: * [Assumption 1: e.g., Testers have basic system knowledge] * [Assumption 2: e.g., Environment will be stable] **Dependencies**: * [Dependency 1: e.g., UAT environment must be ready by [date]] * [Dependency 2: e.g., Test data must be loaded by [date]] --- ## 11. Sign-off ### UAT Completion Sign-off | Name | Role | Signature | Date | | :--- | :--- | :--- | :--- | | [Name] | Business Owner | __________ | __________ | | [Name] | UAT Lead | __________ | __________ | | [Name] | Product Owner | __________ | __________ | | [Name] | Project Manager | __________ | __________ | **Comments**: [Any conditions, caveats, or notes associated with sign-off] --- ## 12. Appendices ### Appendix A: Test Case Details [Link to full test case document or test management tool] ### Appendix B: Defect Log [Link to defect tracking system or spreadsheet] ### Appendix C: Training Materials [Links to user guides, training videos, or documentation] ### Appendix D: Test Data Reference [Links to test data sets or data preparation scripts]

Latest Blog Posts

MCP directory API

We provide all the information about MCP servers via our MCP API.

curl -X GET 'https://glama.ai/api/mcp/v1/servers/DauQuangThanh/sso-mcp-server'

If you have feedback or need assistance with the MCP directory API, please join our Discord server