Control 2.5: Testing, Validation, and Quality Assurance - Troubleshooting
This playbook provides troubleshooting guidance for Control 2.5.
Common Issues and Solutions
| Issue | Symptoms | Root Cause | Solution |
|---|---|---|---|
| Test environment mismatch | Tests pass in test but fail in production | Configuration drift | Compare and sync environment settings |
| Intermittent test failures | Same test passes sometimes, fails other times | Race conditions | Add wait times, retry logic |
| UAT delays | Business users not completing UAT | Unclear expectations | Provide clear scenarios, schedule dedicated time |
| Golden dataset outdated | Regression tests failing on valid responses | Knowledge sources updated | Update golden dataset entries |
| Hallucination rate increase | Groundedness scores declining | Knowledge source gaps | Review and expand knowledge base |
Detailed Troubleshooting
Issue 1: Test Environment Not Matching Production
Symptoms: Tests pass in test but fail in production
Resolution:
- Compare environment configurations:
- DLP policies
- Security roles
- Connection references
-
Environment variables
-
Verify DLP policies match:
- Export DLP from both environments
- Compare connector allowances
-
Sync any differences
-
Check data source connectivity:
- Verify test environment can reach same data sources
-
Check authentication methods match
-
Review security role differences:
- Compare role assignments
-
Verify service accounts exist in both
-
Sync solution versions:
- Export from production
- Import to test environment
Issue 2: Automated Tests Failing Intermittently
Symptoms: Same test passes sometimes, fails other times
Resolution:
-
Add appropriate wait times:
# Add delay between requests Start-Sleep -Milliseconds 500 -
Check for race conditions:
- Review async operations
-
Add proper synchronization
-
Review test data dependencies:
- Ensure test data is reset between runs
-
Check for data state dependencies
-
Increase timeout values:
$response = Invoke-RestMethod -Uri $endpoint -TimeoutSec 60 -
Add retry logic:
$maxRetries = 3 for ($i = 0; $i -lt $maxRetries; $i++) { try { $response = Invoke-RestMethod -Uri $endpoint break } catch { if ($i -eq $maxRetries - 1) { throw } Start-Sleep -Seconds 2 } }
Issue 3: UAT Delays
Symptoms: Business users not completing UAT
Resolution:
- Provide clear test scenarios:
- Write step-by-step instructions
- Include expected outcomes
-
Provide example inputs
-
Schedule dedicated UAT time:
- Block calendars for testing
- Remove competing priorities
-
Set clear deadlines
-
Offer testing support:
- Assign QA resource to assist
- Provide FAQ document
-
Hold office hours for questions
-
Simplify test documentation:
- Use screenshots
- Create video walkthroughs
-
Reduce required documentation
-
Set firm deadlines with escalation:
- Communicate consequences of delay
- Escalate to management if needed
Issue 4: Golden Dataset Outdated
Symptoms: Valid agent responses failing regression tests
Resolution:
- Review failed test cases:
- Identify why expected response no longer matches
-
Determine if agent response is actually correct
-
Update golden dataset:
- Modify expected_answer_contains patterns
-
Add new acceptable variations
-
Establish update process:
- Schedule quarterly reviews
- Assign ownership for maintenance
-
Track knowledge source updates
-
Version golden datasets:
- Tag with date and version
- Maintain history of changes
- Document reasons for updates
Issue 5: Hallucination Rate Increasing
Symptoms: Groundedness scores declining, inaccurate responses
Resolution:
- Analyze failed responses:
- Review specific hallucinations
-
Identify patterns or topics
-
Check knowledge sources:
- Verify sources are current
- Look for gaps in coverage
-
Check for conflicting information
-
Review agent prompts:
- Check system prompts for clarity
- Ensure grounding instructions are clear
-
Add constraints if needed
-
Expand knowledge base:
- Add missing topics
- Update outdated content
-
Improve source quality
-
Adjust agent settings:
- Increase grounding strictness
- Add fallback behaviors for uncertain topics
Issue 6: Performance Degradation
Symptoms: Response times exceeding thresholds
Resolution:
- Check concurrent usage:
- Review usage patterns
-
Identify peak times
-
Review knowledge source size:
- Large knowledge bases slow responses
-
Consider chunking or filtering
-
Check connector performance:
- Test individual connectors
-
Look for slow data sources
-
Review agent complexity:
- Complex topic flows increase latency
-
Simplify where possible
-
Monitor environment resources:
- Check Dataverse performance
- Review environment limits
Issue 7: Security Tests Failing
Symptoms: Prompt injection or data leakage detected
Resolution:
- Review agent configuration:
- Check system prompt safeguards
-
Verify content moderation settings
-
Strengthen prompts:
- Add explicit security instructions
-
Include boundary definitions
-
Enable content filtering:
- Configure Azure AI Content Safety
-
Block harmful patterns
-
Test additional attack vectors:
- Expand security test cases
-
Include new injection techniques
-
Document and escalate:
- Log security findings
- Involve security team
Escalation Path
If issues cannot be resolved using this guide:
- Level 1: QA Lead - Test methodology and execution
- Level 2: AI Governance Lead - Policy and standards
- Level 3: Compliance Officer - Regulatory requirements
- Level 4: Microsoft Support - Product-level issues
Related Playbooks
- Portal Walkthrough - Step-by-step portal configuration
- PowerShell Setup - Automation scripts
- Verification & Testing - Test procedures
Updated: January 2026 | Version: v1.2