An AI-assisted coding tool where you define outcomes and pay only for verified code
npm install outcome-clibash
npm install -g outcome-cli
`
This installs the Outcome CLI globally, making the outcome command available system-wide.
$3
- Node.js 18+
- At least one API key (OpenAI, Anthropic, or Google Gemini)
$3
After installation, create a .env file in your project directory or set environment variables:
`bash
Create .env file
echo "OPENAI_API_KEY=sk-proj-your-key-here" > .env
echo "ANTHROPIC_API_KEY=sk-ant-api03-your-key-here" >> .env
echo "GOOGLE_API_KEY=AIzaSy-your-key-here" >> .env
`
Or set environment variables directly in your shell.
Quick Start
Get Outcome running and generate your first verified code feature in under 5 minutes.
$3
`bash
Define an authentication feature outcome
outcome define auth-feature \
--description "Add JWT user authentication with password hashing" \
--criteria tests-pass,lint-clean,builds-successfully \
--max-attempts 3
`
$3
`bash
Multiple agents compete to deliver
outcome run auth-feature \
--models gpt-4o,claude-sonnet \
--live
`
$3
`bash
Deterministic verification
outcome verify auth-feature
Expected output:
ā
Outcome 'auth-feature' PASSED
š Criteria met: tests-pass, lint-clean, builds-successfully
š° Cost: $2.50 (only if all criteria passed)
`
Usage
$3
#### Define an Outcome
`bash
outcome define [options]
Options:
--description Description of the outcome
--criteria Comma-separated verification criteria
Available: tests-pass, lint-clean, builds-successfully,
security-scan, benchmark-passes
--max-attempts Maximum attempts per agent (default: 3)
--timeout Timeout per attempt (default: 300000)
--budget Maximum cost per outcome (default: 10.00)
`
#### Run an Outcome
`bash
outcome run [options]
Options:
--models Comma-separated models to use
Available: gpt-4o, claude-sonnet, claude-opus,
gemini-pro, gemini-flash
--live Show real-time progress
--parallel Number of agents to run in parallel (default: 3)
`
#### Verify an Outcome
`bash
outcome verify
Shows detailed verification results including:
- Which criteria passed/failed
- Test output
- Build status
- Performance benchmarks
`
#### List Outcomes
`bash
outcome list
Shows all defined outcomes with status
`
$3
`bash
Define the outcome
outcome define user-registration \
--description "Implement user registration with email verification" \
--criteria tests-pass,lint-clean,builds-successfully,security-scan \
--max-attempts 5
Run with multiple models
outcome run user-registration \
--models gpt-4o,claude-sonnet,gemini-pro \
--live
Verify the result
outcome verify user-registration
`
$3
- tests-pass: All unit tests pass (Jest, Vitest, etc.)
- lint-clean: Code passes linting (ESLint, Prettier)
- builds-successfully: Project builds without errors
- security-scan: Passes security analysis
- benchmark-passes: Meets performance benchmarks
Architecture
`
outcome-cli/
āāā src/
ā āāā outcomes/ # Outcome definitions and management
ā āāā eval/ # Core evaluation engine
ā ā āāā evaluateOutcome.ts # Binary evaluation logic
ā ā āāā validators.ts # Verification functions
ā ā āāā ai-judge.ts # LLM-as-judge system
ā ā āāā weighted-scorer.ts # DAWS scoring system
ā āāā runtime/ # Multi-model adapters
ā ā āāā openai-adapter.ts
ā ā āāā anthropic-adapter.ts
ā ā āāā google-adapter.ts
ā āāā agents/ # Agent configurations
ā āāā commands/ # CLI command handlers
ā āāā define.ts
ā āāā run.ts
ā āāā verify.ts
`
$3
- Binary Evaluation Engine: Deterministic pass/fail verification
- AI Judge System: LLM-based code quality assessment with caching
- Validator Framework: Extensible validation functions
- Weighted Scoring System: Multi-metric performance evaluation
- Multi-Model Runtime: Unified interface across AI providers
Contributing
We welcome contributions! Areas of focus:
- New verification criteria
- Additional AI model support
- Performance optimizations
- Documentation improvements
- Test coverage
$3
`bash
Install dependencies
npm install
Run tests
npm test
Run linting
npm run lint
Build the project
npm run build
`
$3
1. Create a validator function in src/eval/validators.ts`