Your organization has deployed a firewall, an EDR solution, a SIEM, email filtering, network segmentation, and a vulnerability management program. You have spent six or seven figures on security technology. The question you cannot answer with any of those tools alone is: does any of it actually work against real attacks?
Breach and Attack Simulation (BAS) answers that question by safely executing real attack techniques against your production environment and measuring whether your security controls detect, prevent, or miss them. Unlike penetration testing, which provides a point-in-time snapshot from a human tester, BAS runs continuously and automatically, validating your defenses against a constantly updated library of attack scenarios.
The market has grown rapidly. Gartner included BAS in its 2024 Hype Cycle for Security Operations and projects a 25% CAGR through 2028. But the concept is simple: simulate attacks, measure detection, fix gaps, repeat.
What Is Breach and Attack Simulation?
BAS platforms execute real attack techniques in a controlled, safe manner against your production environment to test whether your security controls respond correctly. A typical BAS simulation might:
- Send a phishing email with a known malicious payload to test whether your email gateway blocks it
- Attempt lateral movement using credential-based techniques to test whether your EDR detects the behavior
- Exfiltrate data over DNS or HTTPS to test whether your DLP and network monitoring detect the data transfer
- Execute a ransomware simulation (without actual encryption) to test whether your endpoint protection stops the execution chain
- Exploit a known CVE against a vulnerable service to test whether your IDS/IPS signatures are current and effective
Each simulation produces a clear pass/fail result: the control either detected the attack or it did not. Over hundreds of simulations mapped to the MITRE ATT&CK framework, you build a quantitative picture of your detection coverage across the entire kill chain.
BAS vs Penetration Testing vs Red Teaming
These three disciplines are complementary, not interchangeable. Understanding the differences helps you invest appropriately in each.
Penetration Testing
- Frequency: Annual or semi-annual
- Approach: Human testers attempt to compromise systems within a defined scope and timeframe
- Strength: Finds complex, chained vulnerabilities that automated tools miss. Tests business logic and human factors.
- Limitation: Point-in-time snapshot. A configuration change the day after the test can introduce new gaps that are not caught until next year.
Red Teaming
- Frequency: Annual or ad hoc
- Approach: Adversary simulation with realistic TTPs, often including social engineering, physical access, and supply chain vectors
- Strength: Tests the entire security program, including people and processes, not just technology
- Limitation: Expensive, requires specialized talent, and provides a narrative rather than quantitative metrics
Breach and Attack Simulation
- Frequency: Continuous (daily, weekly, or on-demand)
- Approach: Automated execution of known attack techniques against production controls
- Strength: Continuous validation, quantitative coverage metrics, immediate visibility into control drift
- Limitation: Tests known techniques from a library. Cannot discover novel attack paths or test human factors.
The ideal program uses all three: BAS for continuous baseline validation, penetration testing for deep technical assessment, and red teaming for realistic adversary simulation. Each fills gaps the others cannot.
MITRE ATT&CK Mapping: The Foundation of BAS
The MITRE ATT&CK framework provides the taxonomy that makes BAS results actionable. ATT&CK documents real-world adversary tactics and techniques in a structured matrix, from initial access through execution, persistence, privilege escalation, defense evasion, credential access, discovery, lateral movement, collection, command and control, exfiltration, and impact.
A well-designed BAS program maps every simulation to a specific ATT&CK technique. This produces a coverage heatmap that shows exactly where your defenses are strong and where gaps exist:
- Green (detected and blocked): Your controls successfully identified and stopped the technique. Validate these quarterly to ensure detection remains effective.
- Yellow (detected but not blocked): Your SIEM or EDR generated an alert, but the attack was not automatically prevented. This indicates a detection engineering or response automation gap.
- Red (not detected): The attack technique executed successfully without triggering any alert. This is a blind spot that an attacker could exploit.
- Gray (not tested): No simulation exists for this technique in your BAS library, or the technique is not applicable to your environment.
The heatmap drives prioritized remediation. Red cells adjacent to techniques used by threat groups that target your industry should be addressed immediately. Red cells for techniques that no known adversary is using can be addressed in the normal remediation cycle.
Building a Continuous Validation Program
Phase 1: Baseline Assessment (Weeks 1-4)
Run a comprehensive BAS assessment against your current security stack. This establishes your baseline detection coverage across the ATT&CK matrix. Expect the results to be sobering. Most organizations discover that their security controls detect fewer than 40% of common attack techniques on the first run.
- Deploy BAS agents or sensors in each network segment you want to test
- Configure simulations for the ATT&CK techniques most relevant to your threat profile
- Run the full simulation library and record results
- Generate the initial ATT&CK coverage heatmap
Phase 2: Detection Engineering (Weeks 5-12)
Address the gaps identified in Phase 1. For each red cell in your ATT&CK heatmap:
- Determine if the detection is possible with your current tooling. Some techniques require specific sensor coverage (e.g., kernel-level monitoring for certain persistence mechanisms).
- Write or tune detection rules. Create SIEM correlation rules, EDR behavioral detections, or network signatures that target the specific technique.
- Re-run the simulation to validate that the new detection works. This is the BAS feedback loop: simulate, detect, tune, re-simulate.
- Document the detection logic so it can be maintained and updated as the technique evolves.
Phase 3: Continuous Validation (Ongoing)
Schedule automated BAS runs on a recurring basis. Recommended cadences:
- Weekly: Run your core simulation set against critical infrastructure. This catches detection drift from configuration changes, signature expirations, or tool updates.
- Monthly: Run the full simulation library including newly added attack techniques. BAS vendors update their libraries regularly as new TTPs are documented.
- On-demand: Run targeted simulations whenever you deploy a new security tool, change a detection rule, or learn about a new threat targeting your industry.
- Post-incident: After any security incident, run simulations for the specific techniques used by the attacker to validate that your defenses now detect and prevent them.
BAS Tools and Platforms
The BAS market includes commercial platforms and open-source options at various price points:
Commercial Platforms
- SafeBreach: Enterprise BAS with extensive ATT&CK coverage, cloud simulation, and integration with major SIEM/SOAR platforms
- AttackIQ: BAS platform built around the ATT&CK framework with a free community edition (AttackIQ Flex) for smaller organizations
- Cymulate: Cloud-delivered BAS covering email, web, network, and endpoint attack vectors with executive-level reporting
- Picus Security: BAS with integrated mitigation recommendations, mapping simulated attacks to specific detection rule improvements
Open-Source Options
- Atomic Red Team: A library of small, focused tests mapped to ATT&CK techniques. Not a full BAS platform, but an excellent starting point for teams that want to build their own validation program.
- Caldera: MITRE's own adversary emulation platform. More complex than Atomic Red Team but provides automated multi-step attack scenarios.
- BASzy AI: Open-source BAS platform with 35+ attack modules mapped to MITRE ATT&CK. Includes AI-powered attack chain planning that automatically selects relevant techniques based on your environment. Designed for teams that want BAS capabilities without enterprise platform costs.
Measuring BAS ROI
BAS programs need to demonstrate value to justify their cost. The metrics that matter:
Detection Coverage Score
The percentage of simulated ATT&CK techniques that your security controls detect. Track this over time. A program that moves from 35% detection coverage to 72% over 12 months has quantifiable proof that security investments are working.
Mean Time to Detection (MTTD) Improvement
BAS-driven detection engineering typically reduces MTTD because it identifies and fixes detection gaps proactively. Measure MTTD before and after BAS-informed tuning to quantify the improvement.
Control Drift Rate
How often do previously-passing simulations start failing? A high drift rate indicates that operational changes (tool updates, configuration changes, staff turnover) are degrading your security posture. BAS makes this drift visible before it is exploited.
Cost Per Detection Gap Closed
Divide your total BAS program cost by the number of detection gaps identified and remediated. Compare this to the cost of discovering the same gaps through a penetration test or, worse, through an actual incident.
Integrating BAS with Vulnerability Management
BAS and vulnerability management are complementary disciplines that become more powerful when integrated:
- Validate vulnerability prioritization: Run BAS simulations that exploit specific CVEs in your environment. A CVE that your scanner rates as critical but that your EDR blocks in simulation may be lower priority than a medium-rated CVE that sails through your defenses undetected.
- Test compensating controls: When a vulnerability cannot be patched immediately, deploy a compensating control (network rule, WAF rule, EDR behavioral detection) and validate it with a BAS simulation that attempts to exploit the specific CVE.
- Prioritize detection engineering: Use vulnerability data to focus BAS simulations on the techniques most likely to be used against your specific vulnerabilities. If your environment has CVEs that enable initial access via T1190 (Exploit Public-Facing Application), run BAS simulations for that technique first.
- Unified reporting: Combine vulnerability posture (what could be exploited) with BAS results (what would succeed) for a complete risk picture. This is the difference between "we have 3,000 vulnerabilities" and "we have 3,000 vulnerabilities, and our controls stop 94% of the attack techniques that would exploit them."
Common BAS Implementation Mistakes
- Running BAS without a remediation process: BAS findings without follow-through are just expensive reports. Every detection gap identified by BAS should enter a remediation workflow with ownership, deadlines, and validation.
- Testing only the easy stuff: Running email phishing simulations and declaring success ignores the entire kill chain after initial access. BAS value comes from testing lateral movement, privilege escalation, data exfiltration, and command and control techniques.
- Ignoring environmental context: Not all ATT&CK techniques are relevant to your environment. Focus simulations on the tactics and techniques used by threat groups that target your industry, as documented in MITRE ATT&CK threat group profiles.
- Running BAS in isolation: BAS results should inform vulnerability management priorities, detection engineering backlogs, and security architecture decisions. If BAS operates as a standalone program disconnected from these functions, its impact is limited.
- Measuring the wrong things: "We ran 500 simulations" is an activity metric, not an outcome metric. "We improved ATT&CK detection coverage from 38% to 67%" is an outcome metric that demonstrates security improvement.
The Bottom Line
Breach and Attack Simulation fills the critical gap between knowing what could be exploited (vulnerability management) and knowing what would succeed (validation). It transforms security from an assumption-based discipline into an evidence-based one. You stop saying "we think our EDR would catch that" and start saying "we tested it last Tuesday and it did."
The organizations that run continuous BAS programs make better vulnerability prioritization decisions, maintain higher detection coverage, and catch control drift before it becomes a breach. The organizations that do not are relying on an annual penetration test and hope. In 2026, hope is not a security strategy.
Whether you start with an open-source tool like BASzy AI and Atomic Red Team or invest in a commercial platform, the important thing is to start validating. Every week you go without testing your defenses is a week an attacker might be testing them for you.