EU AI Act enforcement begins August 2, 2026 — Are you ready?
← Back to Case Studies

CrewAI Governance Audit

CrewAI scores 13/100 on enforcement posture -- the lowest in our audit portfolio. The leading multi-agent framework has zero test files at root level, 56 potential secrets, and no AI agent instructions, creating a governance gap in the very infrastructure designed to orchestrate AI agents.

Overall Score: 13/100 (Grade: F)

10/100
Enforcement Maturity
Grade: F
10/100
Context Hygiene
Grade: F
22/100
Automation Readiness
Grade: D

Executive Summary

CrewAI is the leading multi-agent AI framework with 25,000+ GitHub stars, enabling enterprises to build autonomous AI agent teams. It has rapidly become the default choice for organizations deploying multi-agent architectures in production.

The irony is stark: a framework designed to orchestrate AI agents scores F (13/100) on the governance measures needed to govern those same agents. Zero test files at root level, 56 potential hardcoded secrets, and no CLAUDE.md means AI agents building on CrewAI have no structural guardrails. Governance gaps here cascade into every system built on top of it.

Enforcement Ladder Distribution

L5 - Hooks0 found

No automated enforcement before commits or tool use

L4 - Tests0 at root*

Tests may exist in lib packages but not discovered at root level

L3 - Templates11 workflows

Moderate CI pipeline with GitHub Actions automation

L2 - Prose0 rules

No CLAUDE.md or agent-specific instruction files

L1 - ConversationDefault

Default mode for all interactions

Diagnosis: CrewAI has the weakest enforcement posture in our audit portfolio. The only structural enforcement comes from 11 GitHub Actions workflows at L3. For a framework whose purpose is orchestrating autonomous AI agents, this absence of self-governance is both ironic and concerning. The agents CrewAI orchestrates have more structural guardrails than CrewAI's own development process.

Critical Gaps Found

1. No Hook Enforcement [CRITICAL]

CrewAI has no pre-commit hooks or Claude Code hooks. AI agents can modify any file in the framework without structural gatekeeping. Security-critical agent orchestration logic and tool-use pathways have no modification guards.

2. No Test Coverage [CRITICAL]

Zero test files detected at root level. While tests may exist in individual library packages, no unified test command validates the entire framework. Contributors have no clear testing contract for a framework handling autonomous AI decision-making.

3. Potential Hardcoded Secrets (56) [CRITICAL]

56 instances of potential hardcoded secrets detected -- the highest count in our audit portfolio. No automated secret scanning in CI. API keys, tokens, or credentials may be embedded in source files with no convention for test-only credentials.

4. No CLAUDE.md [HIGH]

No CLAUDE.md or equivalent AI agent instruction file. For a multi-agent framework, this is especially damaging -- AI agents building on or contributing to CrewAI have zero project-specific context, no architectural guardrails, and no knowledge of framework conventions.

5. Extreme TODO Debt (13,838) [MEDIUM]

13,838 TODO/FIXME/HACK markers detected -- an extraordinarily high count likely inflated by markdown documentation in crewai-tools. No systematic process for converting TODOs to actionable work items. AI agents may attempt incorrect "fixes" at scale.

EU AI Act Compliance Mapping

CrewAI is not itself a high-risk AI system, but it is the infrastructure on which autonomous AI agent teams are built. Organizations deploying CrewAI-orchestrated agents in regulated contexts inherit CrewAI's governance gaps directly. As agent infrastructure, CrewAI's compliance posture is multiplied across every system built on it.

Article 9: Risk Management System

RequirementReadiness
9(2)(a) Risk identification5%
9(2)(b) Risk evaluation5%
9(2)(d) Risk management measures10%
9(6) Testing for risk management10%
9(7) Lifecycle risk management5%

Article 15: Accuracy, Robustness and Cybersecurity

RequirementReadiness
15(1) Accuracy levels10%
15(2) Error resilience10%
15(3) Manipulation robustness5%
15(4) Cybersecurity5%

Article 17: Quality Management System

RequirementReadiness
17(1)(a) Compliance strategy5%
17(1)(b) Design/development procedures15%
17(1)(c) Test/validation procedures10%
17(1)(g) Post-market monitoring0%
Overall EU AI Act Readiness: ~12%

This is the lowest compliance readiness in our audit portfolio, and it is especially concerning for a framework that serves as the orchestration layer for autonomous AI agents. Every agent system built on CrewAI inherits these compliance gaps.

Recommendations

Immediate (Week 1)

  1. Create CLAUDE.md with agent architecture overview, core module boundaries, and critical enforcement rules -- 1 hour effort, foundational for all AI-assisted development
  2. Add secret scanning to CI pipeline (truffleHog or detect-secrets) and audit all 56 potential secrets -- 2 hours effort
  3. Add 3 pre-commit hooks for agent orchestration module guards, secret scanning, and test requirements -- 2 hours effort

Short-term (Month 1)

  1. Deploy L5 enforcement hooks for security-critical agent orchestration paths
  2. Create unified test orchestration with root-level runner across all packages
  3. Implement TODO triage to separate documentation artifacts from genuine debt across 13,838 markers

Strategic (Quarter)

  1. Build enforcement ladder documentation mapping to EU AI Act requirements
  2. Establish violation tracking across contributor AI tool usage
  3. Autoresearch optimization -- auto-tune enforcement rules based on violation patterns

Appendix: Raw Scan Data

0*
Test Files
747
Source Files
11
GitHub Actions
56
Potential Secrets
13,838
TODO/FIXME
152
Dead Code Markers
0
CLAUDE.md Files
0
L5 Hooks
0
Doc Files

Want this analysis for your codebase?

Get the same structural governance audit -- risk classification, violation scan, and enforcement recommendations.

Request a Free Audit
This governance audit was generated by walseth.ai using automated enforcement posture scanning. The findings are based on static analysis of the repository structure, configuration files, and code patterns -- no code was executed during the audit.

Get Your Free AI Governance Audit

Submit your repository and receive a structural governance assessment -- risk classification, violation scan, and enforcement recommendations. No cost, no commitment.

Request Free Audit