22 Self-Learning AI Agents - Build Institutional Knowledge That Compounds Over Time - Production-Ready Orchestration Framework
npm install equilateral-agents-open-core22 self-learning AI agents. Build institutional knowledge that compounds over time. MIT licensed.
Transform your AI coding assistant into a learning system that gets smarter with every mistake you make (and prevents you from making it again).




---
Your Codebase Learns From Its Mistakes.
v2.5.0 introduces complete methodology for building institutional knowledge through standards:
``bashWeek 1: Run workflows, see what breaks
npm run workflow:security
npm run workflow:quality
What's Included:
- π Complete Methodology - BUILDING_YOUR_STANDARDS.md, PAIN_TO_PATTERN.md, KNOWLEDGE_HARVEST.md
- π Example Standards - 6 real standards with actual incident costs in
.standards-local-template/
- π― CLAUDE.md Template - Tell your AI assistant to check standards before every change
- π Knowledge Synthesis Flywheel - Execute β Learn β Document β Prevent β Repeat
- π Case Studies - HoneyDoList.vip: 38-40 hours to production SaaS, Agent Orchestration FrameworkThe Value:
- Greenfield projects: Start with best practices, build standards as you learn domain
- Brownfield codebases: Systematically document problems, prevent repeating mistakes
- Growing teams: New developers learn from team's past pain, onboard faster
- Compounding knowledge: Every incident becomes institutional memory
See RELEASE_NOTES_v2.5.0.md for complete details.
---
Why EquilateralAgents?
$3
Traditional development:
- β Same security bugs discovered 3+ times
- β N+1 query performance issues in every new feature
- β Production incidents from patterns you've seen before
- β New developers repeat mistakes the team already solved
- β No institutional memory - knowledge lives in people's heads
$3
EquilateralAgents creates a feedback loop:
`
1. Execute Workflows (agents scan your code)
β
2. Agent Memory (tracks what worked, what failed)
β
3. Knowledge Harvest (extract patterns weekly)
β
4. Create Standards (document "What Happened, The Cost, The Rule")
β
5. Enforce Standards (AI checks before changes, agents validate)
β
6. Fewer Incidents (prevent repeating mistakes)
β
[Loop back to step 1]
`Result: Your codebase gets smarter over time. Mistakes happen once, not repeatedly.
---
Perfect For
$3
Start right from day 1:
- Security scanning before first commit
- Quality gates before bad patterns take root
- Document decisions as you make them
- Build standards library alongside code
Example journey:
- Week 1: Run security/quality workflows, create first standards
- Month 1: 10+ standards covering your specific domain
- Month 3: New feature? Check standards first. AI references them automatically.
$3
Fix systematically, not randomly:
- Agents identify patterns across entire codebase
- Document each fix as a standard (prevent recurrence)
- Gradually eliminate entire classes of bugs
- Track progress: incidents per month going down
Example journey:
- Week 1: Security scan finds 50 issues. Fix 10, document pattern.
- Month 2: Similar issue caught by agent during PR. Standard working.
- Month 6: That entire class of bugs eliminated from codebase.
Real results:
- Production incidents: 8/quarter β 1/quarter (87% reduction)
- Debug time: 4 hours/incident β 0 (caught in PR review)
- ROI: One prevented outage pays for entire year of standards work
---
Quick Start
$3
`bash
Clone repository
git clone https://github.com/Equilateral-AI/equilateral-agents-open-core.git
cd equilateral-agents-open-coreInstall dependencies (zero config - works immediately)
npm installRun first workflow
npm run workflow:security
`No database setup. No API keys. No configuration files. Works immediately.
$3
- [ ] Day 1: Run security and quality workflows on your codebase
- [ ] Day 2: Review
.equilateral/workflow-history.json - what did agents find?
- [ ] Day 3: Copy .standards-local-template/ to .standards-local/
- [ ] Day 4: Create your first standard from most painful issue agents found
- [ ] Day 5: Update .claude/CLAUDE.md to reference your new standardSee BUILDING_YOUR_STANDARDS.md for complete Week 1 β Year 3 roadmap.
---
What's Included
$3
Infrastructure Core (3)
- AgentClassifier - Task routing and complexity analysis
- AgentMemoryManager - Context and state management
- AgentFactoryAgent - Self-bootstrapping agent generation
Development (6)
- CodeAnalyzerAgent - Static analysis and metrics
- CodeGeneratorAgent - Pattern-based code generation
- TestOrchestrationAgent - Multi-framework test execution
- DeploymentValidationAgent - Pre-deployment validation
- TestAgent - UI testing with intelligent element remapping
- UIUXSpecialistAgent - Design consistency and accessibility
Quality Assurance (5)
- AuditorAgent - Standards compliance validation
- CodeReviewAgent - Best practice enforcement
- BackendAuditorAgent - Backend-specific standards
- FrontendAuditorAgent - Frontend-specific standards
- TemplateValidationAgent - IaC template validation
Security (4)
- SecurityScannerAgent - Vulnerability scanning
- SecurityReviewerAgent - Security posture assessment
- SecurityVulnerabilityAgent - Common security issue detection
- ComplianceCheckAgent - Basic compliance validation
Infrastructure (4)
- DeploymentAgent - Deployment automation
- ResourceOptimizationAgent - Cloud resource analysis
- ConfigurationManagementAgent - IaC configuration patterns
- MonitoringOrchestrationAgent - Observability best practices
See AGENT_INVENTORY.md for complete capabilities.
$3
Documentation:
- BUILDING_YOUR_STANDARDS.md - Week 1 β Year 3 roadmap
- PAIN_TO_PATTERN.md - "What Happened, The Cost, The Rule" methodology
- KNOWLEDGE_HARVEST.md - Daily/weekly pattern extraction process
- .claude/CLAUDE.md - Template for AI assistant integration
Example Standards (.standards-local-template/):
- Security: Credential scanning, input validation, auth & access control
- Architecture: Error-first design patterns
- Performance: Database query optimization, N+1 prevention
- Testing: Integration tests without mocks
The Difference:
- Open-core: Methodology + templates + 22 agents (teach you to fish)
- Commercial: 138+ battle-tested standards + 62 agents (give you 138 fish already caught)
$3
The project-object skill is included by default, closing the loop between standards and enforcement. Without it, YAML standards sit in a directory. With it, they're actively injected into every AI session.
How it works:
1. On session start: The skill reads
~/.project-object/{project}/context.md and injects prior session context (decisions, patterns, corrections) into the agent's system prompt
2. Standards loading: The standards-loader.js scans .standards/yaml/*.yaml, extracts rules by action type, and injects them as [REQUIRE]/[AVOID]/[PREFER] enforcement directives
3. During session: Agent follows established context and enforces standards automatically
4. On session end: Agent harvests new decisions, patterns, and corrections from the conversation transcriptWhat agents get:
- Session memory: Decisions, patterns, corrections, and notes persist between sessions
- Standards injection: YAML standards from
.standards/yaml/ loaded and enforced automatically
- Cross-platform sync: Context syncs to Cursor, Codex, Windsurf, and other AI tools via project-object syncExample injected standards (from your
.standards/yaml/ files):
`
[REQUIRE] Fail fast and loud -- make failures obvious and immediate
[REQUIRE] Use environment variables with {{resolve:ssm:param}} in SAM templates
[AVOID] Return mock data or fallback values from production code on failure
[AVOID] Use connection pools in Lambda -- Lambda handles one request at a time
[PREFER] ARM64 architecture for Lambda functions (20% cost savings)
`Files:
-
.agents/skills/project-object/ - Full skill (SKILL.md, scripts, references)
- .claude/skills/project-object - Symlink for Claude Code auto-discovery
- Also available standalone: npx skills add Equilateral-AI/project-object-skillFor adaptive learning (automatic correction detection, invariant promotion), see MindMeld.
$3
`bash
npm run workflow:security # Multi-layer security assessment
npm run workflow:quality # Code quality analysis (0-100 score)
npm run workflow:deploy # Deployment validation
npm run workflow:fullstack # Full-stack development workflow
npm run workflow:infrastructure # Infrastructure validation
`See workflows/README.md for details.
$3
Agents automatically:
- Track last 100 executions
- Identify success/failure patterns
- Suggest optimizations
- Improve recommendations over time
You manually:
- Review agent memory weekly (
npm run memory:stats)
- Extract patterns ("this error happened 3+ times")
- Create standards (document "What Happened, The Cost, The Rule")
- Update .claude/CLAUDE.md (AI checks standards before changes)Commercial upgrade:
- Librarian agent automates knowledge harvest
- Pattern recognition ML across projects
- Cross-enterprise learning (anonymized)
---
Three-Tier Standards System
EquilateralAgents uses a hierarchical standards approach:
$3
EquilateralAgents Open Standards - Universal principlesCore principles:
- No mocks in production code (test real dependencies)
- Error-first design (design errors before happy paths)
- Cost-conscious infrastructure (estimate before deploying)
- Explicit over implicit (obvious code beats clever code)
$3
Community Patterns - Battle-tested patterns (optional)Contributed by users:
- Agent coordination patterns
- Real-world examples
- Custom workflows
- Integration patterns
Your standards can graduate here after 3+ months of successful use.
$3
Your Team's Standards - Project-specific conventions (git-ignored or private repo)Built from your experience:
- Document incidents as they happen
- "What Happened, The Cost, The Rule" format
- Prevent repeating your specific mistakes
- Your institutional knowledge
$3
`bash
Clone with official standards
git clone --recurse-submodules https://github.com/Equilateral-AI/equilateral-agents-open-core.gitAdd community standards (optional)
git submodule add https://github.com/Equilateral-AI/EquilateralAgents-Community-Standards.git .standards-communityCreate your local standards
cp -r .standards-local-template .standards-local
`---
Integration with AI Assistants
$3
`bash
/plugin marketplace add Equilateral-AI/equilateral-agents-open-core
/plugin install equilateral-agents-open-coreAvailable slash commands
/ea:security-review # Multi-layer security assessment
/ea:code-quality # Code analysis with quality scoring
/ea:memory # View agent learning statistics
/ea:list # See all available workflows
`$3
EquilateralAgents includes
.claude/CLAUDE.md that tells your AI assistant:`markdown
Before Every Code Change:
1. CHECK STANDARDS FIRST
- Read
.standards/ for universal principles
- Check .standards-community/ for proven patterns
- Review .standards-local/ for team conventions2. DESIGN ERRORS FIRST
- What can go wrong? How will it fail?
3. VALIDATE BEFORE COMMIT
- Run relevant agents (security, quality, tests)
- Check agent memory for similar past failures
`Result: AI automatically references your standards, preventing mistakes before code is written.
---
Background Execution
The Pattern: "Dispatch teams in background, execute next todo list tasks"
`javascript
const AgentOrchestrator = require('./equilateral-core/AgentOrchestrator');const orchestrator = new AgentOrchestrator({ enableBackground: true });
await orchestrator.start();
// Dispatch teams in background
const securityTask = orchestrator.executeWorkflowBackground('security-review', {
projectPath: process.cwd()
});
const qualityTask = orchestrator.executeWorkflowBackground('code-quality', {
projectPath: process.cwd()
});
// Continue working on next todo while agents run
await workOnNextTodoListItems();
// Check results when ready
const securityResults = await securityTask.getResult();
const qualityResults = await qualityTask.getResult();
`See BACKGROUND_EXECUTION.md for complete API.
---
Knowledge Synthesis Flywheel
The system that makes your codebase smarter over time:
$3
1. Run workflows on your actual codebase
2. Review findings - agents will find issues
3. Document first pain - create 3-5 standards from most painful issues
4. Update CLAUDE.md - tell AI to check your new standards
$3
1. Weekly review: Check
npm run memory:stats
2. Identify patterns: What failed 3+ times?
3. Create standards: Use "What Happened, The Cost, The Rule" format
4. Measure impact: Track prevented incidents$3
1. Pre-commit hooks: Run agents before every commit
2. CI/CD integration: Block PRs with critical violations
3. Team training: Share standards library, explain why each exists
4. Celebrate wins: Count prevented incidents, estimate cost savings
$3
- 30-50 standards covering most common mistakes
- 87% reduction in production incidents (real data from commercial users)
- 40% faster velocity (less debugging, more building)
- Faster onboarding (new devs learn from documented pain)
$3
- Standards library stabilizes (most patterns documented)
- Focus shifts to enforcement and refinement
- Consider contributing valuable patterns to community
- Explore commercial upgrade for specialized needs
The Goal: Every mistake happens once, gets documented, never repeats.
---
Real Results
$3
Background: New SaaS application, 3 developers, 6 months
Week 1:
- Ran security/quality workflows
- Found 0 issues (greenfield), created 5 standards for domain patterns
- Set up pre-commit hooks
Month 3:
- 15 standards documented (authentication, data validation, API patterns)
- 0 production incidents (agents caught issues in PR review)
Month 6:
- 25 standards, mature workflow
- New developer onboarded in 2 days (read standards, understood decisions)
- Security audit: 95/100 score
$3
Background: Legacy Node.js app, 50k LOC, 5 years old, 8 developers
Week 1:
- SecurityScannerAgent found 47 issues
- BackendAuditorAgent found 30 N+1 queries
- Created first 3 standards from most painful patterns
Month 2:
- Fixed 15 issues, documented patterns as standards
- Agents started catching similar issues in new code
- Prevented 8 incidents (same patterns caught in PR review)
Month 6:
- 35 standards, entire classes of bugs eliminated
- Production incidents: 8/quarter β 1/quarter (87% reduction)
- Debug time per incident: 4 hours β 0 (caught before merge)
Month 12:
- 50+ standards, knowledge library mature
- Team velocity up 40% (less firefighting, more building)
- ROI: One prevented outage paid for entire year of work
---
Open-Core vs Commercial
$3
β
22 production-ready agents - Everything needed to start
β
Complete methodology - Build your own standards library
β
Self-learning system - Agent memory, pattern recognition
β
Background execution - Parallel workflow execution
β
Example standards - 6 templates showing proper format
β
Community contribution - Contribute & benefit from shared knowledge
β
This entire methodology - Teach you to fish
Perfect for:
- Startups and small teams
- Learning the methodology
- Building your first 50 standards
- Contributing to community
$3
β 62 specialized agents (40+ beyond open-core)
β 138+ battle-tested standards (from years of enterprise pain)
β GDPR/HIPAA/SOC2 compliance (specialized domain expertise)
β Librarian agent (automated knowledge harvest)
β Pattern recognition ML (cross-enterprise learning)
β Multi-account AWS (Control Tower integration)
β Advanced security (STRIDE threat modeling, penetration testing)
β Cost intelligence (ML-based predictions)
Perfect for:
- Enterprises with compliance requirements
- Teams that need 138+ standards immediately (skip 2 years of learning)
- Multi-cloud deployments
- Cross-project pattern recognition
$3
Open-core teaches you to fish (methodology + tools)
Commercial gives you 138 fish already caught (battle-tested standards + automation)
$3
Start with open-core. Build your
.standards-local/. Upgrade when you need:
- Specialized compliance (GDPR, HIPAA)
- 138+ pre-built standards (skip years of learning)
- ML-based cost predictions
- Automated knowledge harvest (librarian agent)
- Cross-enterprise pattern recognitionContact: info@happyhippo.ai
---
Contributing
Contributions welcome! See CONTRIBUTING.md for guidelines.
Found a universal pattern? Submit to EquilateralAgents Open Standards
Built something useful? Share with Community Standards
Your battle-tested pattern could help thousands of developers avoid the same mistakes.
---
Security Notice
Important: EquilateralAgents runs with your user account privileges.
Agents can:
- Read/write files in your project
- Execute shell commands
- Access environment variables (API keys, tokens)
- Make network requests
Best Practices:
- Review agent code before running
- Use separate API keys for development
- Run in isolated environments for untrusted workflows
- Monitor agent activity logs in
.equilateral/See SECURITY.md for complete guidelines.
---
Documentation
Case Studies:
- HoneyDoList.vip: Production SaaS in 38-40 Hours - Real-world validation
- Agent Orchestration Framework - FI/FDI decision framework
Methodology Guides:
- BUILDING_YOUR_STANDARDS.md - Week 1 β Year 3 roadmap
- PAIN_TO_PATTERN.md - "What Happened, The Cost, The Rule" methodology
- KNOWLEDGE_HARVEST.md - Daily/weekly pattern extraction
Reference:
- Agent Inventory - All 22 agents with capabilities
- Workflows - Complete workflow guide
- Background Execution - Async API reference
- Plugin Usage - Skills and slash commands
- Protocols - MCP, A2A, WebSockets
Release Notes:
- v2.5.0 Release Notes - Standards methodology
- All Releases - Version history
---
License
MIT License - see LICENSE
Trademarks: EquilateralAgentsβ’ and Equilateral AIβ’ are trademarks of HappyHippo.ai
---
The Bottom Line
Traditional development: Make mistakes repeatedly. Knowledge lives in people's heads. New developers repeat old mistakes.
With EquilateralAgents: Make mistakes once. Document them. Build institutional memory. Your codebase learns.
- Week 1: Run workflows, see what breaks
- Month 2: 10+ standards from your real pain
- Year 1: 30-50 standards preventing entire classes of bugs
- Year 2+: Knowledge compounds, velocity increases, incidents decrease
Your 100th standard represents 100 mistakes you'll never make again.
---
Built by HappyHippo.ai
Ready to start?
`bash
git clone https://github.com/Equilateral-AI/equilateral-agents-open-core.git
cd equilateral-agents-open-core
npm install && npm run workflow:security
``