Skip to main content

Control Status and Evidence Requirements

Learn how control status is determined, what evidence is required, and how to collect and maintain proof of control effectiveness.

John Ozdemir avatar
Written by John Ozdemir
Updated over a month ago

Control status reflects implementation progress and ongoing effectiveness. Understanding status indicators and evidence requirements helps you maintain continuous compliance.

Control Status Explained

DSALTA uses four status levels to indicate control health:

Completed (Green)

The control is fully implemented and verified:

  • All required evidence has been collected

  • Associated tests are passing

  • Policies are documented and approved

  • The owner is assigned and active

  • No outstanding issues

Completed controls contribute to your framework completion percentage and indicate audit readiness for this requirement.

In Progress (Yellow)

The control is being implemented but is not yet complete:

  • Some evidence exists, but gaps remain

  • Tests may be passing, but the evidence is insufficient

  • Policy is drafted but not yet approved

  • Implementation is underway, but not finished

In Progress controls show active work but won't count toward completion until fully implemented.

Needs Attention (Red)

The control has issues requiring immediate action:

  • Previously passing tests are now failing

  • Evidence has expired or become outdated

  • Integration connectivity issues prevent monitoring

  • Critical evidence is missing

Needs Attention indicates drift from compliance or emerging problems that could impact audit outcomes.

No Evidence (Gray)

No implementation activity has occurred:

  • No evidence collected

  • No owner assigned

  • No tests configured

  • No related policies

Not Started controls represent remaining work in your compliance program.

How Status is Determined

DSALTA calculates control status using weighted factors:

Test Results (40%): Are automated tests passing?

Evidence Completeness (35%): Is the required evidence present and current?

Policy Documentation (15%): Are related policies approved?

Manual Verification (10%): Has the owner confirmed implementation?

All factors must be satisfied for the Completed status. Any failing factor triggers In Progress or Needs Attention status.

Evidence Requirements

Evidence proves you've implemented a control and it's working as intended. Each control requires specific evidence types:

Automated Evidence

Collected automatically from integrations:

Identity Provider Evidence:

  • User directory exports

  • MFA enrollment reports

  • Access logs

  • Group membership records

Cloud Infrastructure Evidence:

  • Encryption settings

  • Network configurations

  • Security group rules

  • IAM policies

  • Audit logs

Development Tools Evidence:

  • Code review records

  • Commit histories

  • Vulnerability scan results

  • Deployment logs

Communication Tools Evidence:

  • Security training completion

  • Policy acknowledgments

  • Incident response channels

Automated evidence is collected continuously and updated in real-time as your environment changes.

Manual Evidence

Uploaded by control owners for requirements not covered by integrations:

Documentation:

  • Policy documents

  • Procedures and runbooks

  • Contracts and agreements

  • Certifications and attestations

Screenshots:

  • Configuration settings

  • Dashboard views

  • Security tool outputs

  • System settings

Reports:

  • Vulnerability assessments

  • Penetration test results

  • Risk assessments

  • Audit reports

Records:

  • Training attendance

  • Background check confirmations

  • Access review sign-offs

  • Incident response logs

Evidence Best Practices

Quality Over Quantity

Focus on relevant, clear evidence rather than uploading excessive documentation. Auditors value:

  • Recent evidence (within the last 3-6 months)

  • Clear demonstrations of control effectiveness

  • Well-organized and labeled artifacts

  • Evidence showing continuous operation, not a one-time setup

Proper Labeling

Name evidence files descriptively:

  • βœ… "AWS_Encryption_Settings_2024-12-10.pdf"

  • ❌ "Screenshot.png"

Include dates in filenames to track evidence age.

Regular Updates

Refresh evidence periodically:

  • Quarterly for most controls

  • Monthly for high-risk controls

  • Annually for stable administrative controls

Set reminders in DSALTA to update evidence before it becomes stale.

Complete Coverage

Ensure evidence fully demonstrates the control:

  • Incomplete: Screenshot of MFA enabled for one account

  • Complete: Report showing MFA enabled for all accounts with enrollment dates

Auditors look for comprehensive proof, not partial implementation.

Evidence Collection Workflow

For Automated Evidence:

  1. Connect relevant integration

  2. Allow 15-30 minutes for initial sync

  3. Review collected evidence in control detail

  4. Verify evidence covers all requirements

  5. Monitor integration health to ensure continuous collection

For Manual Evidence:

  1. Open the control detail page

  2. Navigate to the Evidence tab

  3. Click Upload Evidence

  4. Select file(s) from your computer

  5. Add description and evidence type

  6. Associate with specific control requirements

  7. Save and confirm upload

Evidence Retention

DSALTA retains all evidence indefinitely, creating a historical record:

  • Demonstrates continuous compliance over time

  • Provides audit trail of control evolution

  • Shows response to previous audit findings

  • Supports annual recertification

Older evidence remains accessible even after being replaced with updated versions.

Evidence by Control Category

Different control types require different evidence:

Access Control Evidence

  • User directories with role assignments

  • MFA enrollment reports

  • Access review sign-offs

  • Privileged access logs

  • Termination records

Encryption Evidence

  • Encryption configuration screenshots

  • Certificate inventories

  • Key management policies

  • Encryption-at-rest confirmations

  • TLS/SSL test results

Monitoring Evidence

  • Log retention configurations

  • SIEM dashboard screenshots

  • Alert rule definitions

  • Security monitoring reports

  • Incident detection records

Business Continuity Evidence

  • Backup schedules and logs

  • Disaster recovery plans

  • Business continuity test results

  • Recovery time/point objectives

  • Failover documentation

Vendor Management Evidence

  • Vendor security assessments

  • Contract security clauses

  • Vendor risk scores

  • Due diligence reports

  • Vendor monitoring records

Common Evidence Gaps

Avoid these frequent mistakes:

Outdated Evidence: Last year's access review doesn't prove current access is appropriate

Configuration Only: Showing encryption is enabled doesn't prove it's been enabled continuously

Partial Coverage: MFA evidence for some users doesn't satisfy "all users" requirement

Missing Context: Screenshots without explanation of what they demonstrate

Point-in-Time Only: One backup log doesn't prove regular backup schedule

Evidence for Multiple Frameworks

When a control maps to multiple frameworks, the same evidence often satisfies all requirements:

Example: Encryption evidence proving:

  • SOC 2 CC6.7 compliance

  • ISO 27001 A.10.1.1 compliance

  • HIPAA 164.312(a)(2)(iv) compliance

Upload once, apply to all mapped controls automatically.

Organizing Evidence

DSALTA organizes evidence by:

  • Control: All evidence for a specific control

  • Framework: All evidence supporting a framework

  • Date: Chronological evidence collection

  • Type: Automated vs. manual evidence

  • Source: Which integration provided it

This multi-dimensional organization ensures evidence is easily retrievable during audits.

Evidence Review Process

Establish regular evidence reviews:

Monthly: Review critical controls

  • Verify automated evidence is current

  • Check for missing manual evidence

  • Update outdated documentation

Quarterly: Comprehensive evidence audit

  • Review all control evidence

  • Replace stale evidence

  • Fill identified gaps

  • Prepare for potential audits

Annually: Full program review

  • Archive old evidence

  • Update all policies and procedures

  • Refresh all manual evidence

  • Prepare for recertification

Evidence and Audit Readiness

Auditors evaluate evidence quality based on:

Relevance: Does it prove the control exists? Reliability: Is the source trustworthy? Completeness: Does it cover all requirements? Currency: Is it recent enough? Consistency: Does it align with other evidence?

Strong evidence in DSALTA translates to smooth audits with fewer follow-up questions.

Next Steps

To maintain strong control evidence:

  1. Connect integrations to automate evidence collection

  2. Review evidence requirements for each control

  3. Upload manual evidence for gaps

  4. Set quarterly reminders to refresh evidence

  5. Monitor evidence age and replace as needed

Did this answer your question?