1. The Problem: How Do You Know If Your AI Is Trustworthy?
Enterprise AI adoption is accelerating — but trust is not keeping pace. When a CISO is asked "Are our AI systems secure?" by the board, the honest answer is usually "We think so, but we can't quantify it." When a compliance officer needs to demonstrate AI governance for an audit, the evidence is scattered across spreadsheets, Jira tickets, and tribal knowledge.
The fundamental challenge is measurement. Security teams have decades of experience measuring network security (vulnerability counts, patch rates, MTTR), application security (SAST/DAST findings, pen test results), and compliance (control coverage percentages). But AI system trustworthiness? There has been no equivalent metric — until now.
2. Introducing the Trust Score: A Single 0-100 Metric
The ASTRA BASTION Trust Score condenses the complex, multi-dimensional question of "Is this AI system trustworthy?" into a single number between 0 and 100. Like a credit score for AI systems, it provides an immediately understandable measure that can be tracked over time, compared across systems, and reported to stakeholders.
But unlike a simple score, the Trust Score is backed by real, continuously measured data from 6 different data sources across 5 governance pillars. It is not a survey. It is not a self-assessment. It is a computed metric derived from actual system behavior, compliance status, risk posture, and security controls.
Trust Score = Σ (Pillar Score × Pillar Weight)
Where:
Security Pillar × 0.25 (from Gateway + AEGIS data)
Compliance Pillar × 0.25 (from Compliance module assessments)
Risk Pillar × 0.20 (from Risk register + FAIR analysis)
Resilience Pillar × 0.15 (from BAS scenarios + RTO/RPO)
AI Governance × 0.15 (from CyberTwins + policy adherence)
Each pillar score: 0-100, calculated from real operational data.
Final score: 0-100, updated in near real-time.3. The 5 Pillars Deep Dive
Security Pillar (25%)
The Security pillar measures how well your AI systems are protected against attacks. It draws data from two ACL sources: the Gateway ACL (prompt injection detection rates, blocked requests, pipeline effectiveness) and the AEGIS ACL (agent security posture, MCP firewall coverage, kill switch readiness).
- Prompt injection detection rate (what percentage of injection attempts are caught)
- Security control coverage (are all 14 pipeline steps enabled and configured)
- Agent security posture (are agents properly sandboxed and monitored)
- Vulnerability remediation time (how fast are identified issues fixed)
- Kill switch readiness (can you shut down a compromised agent instantly)
Compliance Pillar (25%)
Measures alignment with AI-specific regulations and frameworks. The Compliance ACL provides data from actual framework assessments — not checkboxes, but measured control implementation status.
- EU AI Act compliance coverage (Art. 5-53, with focus on high-risk system requirements)
- NIST AI RMF implementation (24 subcategories across Govern, Map, Measure, Manage)
- Cross-framework control mapping (24 mapped controls between frameworks)
- Evidence freshness (when was compliance evidence last updated)
- Gap count and severity (how many compliance gaps exist and their risk level)
Risk Pillar (20%)
Quantifies the residual risk of AI system operation. Powered by the Risk ACL's Monte Carlo simulations and FAIR quantitative analysis.
- Aggregate residual risk (VaR and CVaR from Monte Carlo simulations)
- Risk treatment effectiveness (how much risk is mitigated by controls)
- Threat landscape coverage (percentage of identified threats with mitigations)
- Risk assessment freshness (when was the last quantitative analysis run)
Resilience Pillar (15%)
Measures the organization's ability to maintain AI operations under adverse conditions and recover from incidents.
- RTO achievement (are AI systems recovering within target time objectives)
- RPO achievement (is data loss within acceptable limits during recovery)
- Scenario coverage (what percentage of failure scenarios have been tested)
- Blast radius containment (can failures be isolated to prevent cascade)
AI Governance Pillar (15%)
Evaluates the maturity of AI governance practices: inventory completeness, policy adherence, and operational oversight.
- AI system inventory completeness (are all AI systems cataloged)
- Policy coverage (do governance policies cover all deployed AI systems)
- Model monitoring (are models being monitored for drift and degradation)
- Human oversight provisions (are appropriate human-in-the-loop controls in place)
4. Real Data from 6 ACL Sources
What makes the Trust Score credible is that every pillar score is calculated from real operational data, not self-reported assessments. ASTRA BASTION's Anti-Corruption Layer (ACL) architecture enables each module to provide standardized data to the Trust Engine without tight coupling.
ComplianceACL → framework_scores, gap_count, evidence_freshness
RiskACL → residual_risk, var_95, cvar_95, treatment_effectiveness
ResilienceACL → rto_achievement, rpo_achievement, scenario_coverage
GatewayACL → detection_rate, pipeline_coverage, blocked_count
AEGISACL → agent_security_score, mcp_coverage, kill_switch_ready
CyberTwinsACL → inventory_completeness, monitoring_coverage
All 6 ACLs queried concurrently via asyncio.gather()
Pillar scores computed from real DB queries (not hardcoded values)
Total computation time: < 200ms for full Trust Score refresh5. The Grade System
Trust Scores are mapped to letter grades that provide immediate, intuitive understanding of an AI system's trustworthiness posture. The grading thresholds are calibrated based on industry benchmarks and regulatory expectations.
A+ 95-100 Exceptional — exceeds all governance requirements
A 90-94 Excellent — strong controls with minor improvements possible
B+ 85-89 Very Good — solid posture with some gaps to address
B 80-84 Good — meets most requirements, improvement areas identified
C 70-79 Adequate — minimum viable governance, significant work needed
D 60-69 Below Average — multiple critical gaps, remediation required
F 0-59 Failing — inadequate governance, immediate action required6. Score Decay: Trust Degrades Over Time
Trust is not static. A system that was secure last month may not be secure today. New vulnerabilities are discovered. Compliance frameworks are updated. Employee AI usage patterns change. The threat landscape evolves continuously.
ASTRA BASTION implements 4 decay models to ensure Trust Scores accurately reflect current reality, not historical snapshots:
- Linear decay: Score decreases at a constant rate when controls are not re-verified. Used for compliance evidence that has a fixed validity period.
- Exponential decay: Score decreases slowly at first, then accelerates. Used for security controls where risk compounds over time without verification.
- Step decay: Score drops by a fixed amount at specific intervals (e.g., quarterly audit deadlines). Used for regulatory compliance milestones.
- Event-driven decay: Score drops immediately when a relevant event occurs (new vulnerability disclosed, policy violation detected, incident reported). Recalculated within 30 seconds via event-driven architecture.
7. How Organizations Use Trust Scores
Trust Scores are not just a dashboard metric. They drive concrete business decisions across the organization:
- Board reporting: CISOs present Trust Score trends to the board as a single KPI for AI governance. A declining score triggers escalation; an improving score demonstrates security investment ROI.
- Vendor selection: When evaluating third-party AI services, organizations use Trust Score criteria as part of their vendor risk assessment. "Must maintain B+ or above" becomes a contractual SLA.
- Regulatory compliance: Trust Score history provides auditable evidence of continuous AI governance — not just point-in-time assessments. EU AI Act Art. 9 risk management requirements are demonstrably met.
- Incident response: After a security incident, the Trust Score provides an objective measure of impact and recovery. "Score dropped from 87 to 62; recovered to 84 within 5 days" is a meaningful metric for post-incident review.
- Development gates: AI systems must achieve a minimum Trust Score (e.g., C or above) before production deployment. This creates a governance gate without slowing down development.
The Trust Score transforms AI governance from a qualitative, subjective process into a quantitative, measurable discipline. For the first time, organizations can answer the question "Are our AI systems trustworthy?" with data, not opinions.