CrewAI Governance Score
The leading multi-agent AI framework scores lowest in our governance portfolio.
Overall Score: 13/100 (Grade: F)
Key Findings
No Hook Enforcement [CRITICAL]
Zero pre-commit hooks or Claude Code hooks. AI agents can modify any file in the framework without structural gatekeeping. Security-critical agent orchestration logic and tool-use pathways have no modification guards.
No Test Coverage at Root Level [CRITICAL]
Zero test files detected at root level. No unified test command validates the entire framework. Contributors have no clear testing contract for a framework handling autonomous AI decision-making.
56 Potential Hardcoded Secrets [CRITICAL]
The highest count in our audit portfolio. No automated secret scanning in CI. API keys, tokens, or credentials may be embedded in source files with no convention for test-only credentials.
Why CrewAI's Governance Score Matters
CrewAI enables enterprises to build autonomous AI agent teams. With 25,000+ GitHub stars, it has rapidly become the default choice for organizations deploying multi-agent architectures in production. The irony is stark: a framework designed to orchestrate AI agents scores F (13/100) on the governance measures needed to govern those same agents.
Governance gaps in agent orchestration infrastructure are especially dangerous because they cascade. Every system built on CrewAI inherits its enforcement posture -- or lack thereof. When the orchestration layer has no structural guardrails, the agents it manages have no foundation to build guardrails upon.
Enforcement Ladder Analysis
CrewAI's enforcement distribution reveals a critical pattern: the only structural enforcement comes from 11 GitHub Actions workflows at L3. No L5 hooks prevent dangerous commits. No L4 tests gate critical paths. No L2 prose (CLAUDE.md) guides AI contributors.
For a framework whose purpose is orchestrating autonomous AI agents, this absence of self-governance creates a compounding risk. The agents CrewAI orchestrates may have more structural guardrails than CrewAI's own development process.
What This Means for Teams Using CrewAI
If your organization deploys CrewAI-orchestrated agents in production, you are inheriting a governance posture that scores 13/100. This does not mean CrewAI is unsafe to use -- it means your team must build the governance layer that CrewAI does not provide. Key areas to address:
- Add pre-commit hooks that validate agent configuration changes before they reach your main branch
- Create a CLAUDE.md for your project that documents how AI agents should interact with CrewAI's API
- Implement secret scanning in your CI pipeline, since CrewAI's own patterns may normalize embedding credentials in code
- Build integration tests that verify agent behavior boundaries, not just functional correctness
EU AI Act Compliance Impact
CrewAI is not itself a high-risk AI system, but it is the infrastructure on which autonomous AI agent teams are built. Organizations deploying CrewAI-orchestrated agents in regulated contexts inherit CrewAI's governance gaps directly. With an estimated 12% EU AI Act readiness -- the lowest in our portfolio -- teams building on CrewAI face significant compliance work before the August 2, 2026 enforcement date.
The most critical gaps are in Article 9 (Risk Management System) and Article 15 (Accuracy, Robustness and Cybersecurity), where readiness scores range from 5% to 10%. For organizations subject to the EU AI Act, these gaps require immediate remediation in your own deployment layer.
Recommendations
Immediate (Week 1): Create CLAUDE.md with agent architecture overview, core module boundaries, and critical enforcement rules (1 hour). Add secret scanning to CI pipeline and audit all 56 potential secrets (2 hours). Add 3 pre-commit hooks for agent orchestration module guards (2 hours).
Short-term (Month 1): Deploy L5 enforcement hooks for security-critical agent orchestration paths. Create unified test orchestration with root-level runner across all packages. Implement TODO triage to separate documentation artifacts from genuine debt.
Strategic (Quarter): Build enforcement ladder documentation mapping to EU AI Act requirements. Establish violation tracking across contributor AI tool usage. Implement autoresearch optimization to auto-tune enforcement rules based on violation patterns.
Raw Scan Data
EU AI Act Readiness
Estimated compliance readiness based on enforcement posture, documentation, and automated quality controls. EU AI Act enforcement begins August 2, 2026.
See how your project compares
Run our free governance scanner on your own repository and get an instant enforcement posture score.
Scan Your Repository