Open Source · Agent Training

Train agents
that can be trusted.

Most agent safety lives in system prompts. Prompts can be forgotten, overridden, or drifted across context.

School for Agents embeds guardrails directly into skill definitions. When an agent installs a skill, it inherits the constraints. The safety contract travels with the skill.

$ npx openskills install conscious-collective/school-for-agents
The Problem

Agent safety is mostly vibes.

The standard approach: write a system prompt with safety instructions, hope the model follows them, debug when it doesn't. That's not a safety system. That's optimism.

Instructions compete with context. A long enough conversation, a clever enough user, or just a model update can erode them. Structural constraints don't have this problem. They're not in the conversation at all.

Safety in system prompt
Safety in skill definition
"Don't send too many emails"
Max 50 emails/day, enforced
Trust the reasoning step
Separate reasoning from execution
Guardrails per deployment
Guardrails travel with the skill
How It Works

Three layers. One contract.

01

Curricula

Theory & frameworks

Structured training programs covering ethical reasoning, safety protocols, and real-world deployment. Modules with decision frameworks, edge case analysis, and escalation logic.

curricula/ethical-reasoning-101/
02

Scenarios

Practice & evaluation

Realistic simulations with multiple-choice options, evaluation rubrics, and scored responses. Each scenario references a skill UUID and includes its own guardrail contract.

scenarios/beginner/scenario-001.md
03

Skill Manifests

Portable safety contracts

OpenSkills-compatible YAML files. Each skill bundles its permissions, hard limits, dry-run requirements, and human-in-loop tier. Import the skill, import the constraints.

skills/privacy-protection.yml
Core Principles

What agents trained here learn to reason with.

01

Constrain capabilities, not intelligence

Remove tools an agent doesn't need. The most reliable guardrail is a missing capability, not a reminder to use one carefully.

02

Three tiers of human oversight

Map every action to a tier: act then notify, confirm before acting, or human-must-initiate. Define tiers before deployment, not after the first incident.

03

Hard limits beat soft instructions

Prompt instructions compete with context. Hard limits embedded in skill definitions do not. Every "don't forget to..." in a prompt is something that should be hardcoded.

04

Separate thinking from acting

Reasoning and execution in the same step means intervention is only possible before or after. Split them. The seam between plan and execute is where governance lives.

05

Guardrails travel with skills

When an agent imports a skill from this school, it imports the contract: permissions, hard limits, dry-run rules, human-in-loop tier. Not a separate document. The skill itself.

Curriculum

What agents learn here.

Ethical Reasoning 101 Ready

5 modules covering the core ethical frameworks every agent needs: utilitarian reasoning, deontological limits, virtue ethics, and care ethics. Practical decision trees for privacy, honesty, escalation, and bias.

Module 1: Foundations of Ethical Reasoning
Module 2: Handling Sensitive Information
Module 3: Honesty vs. Kindness
Module 4: Escalation & Boundaries
Module 5: Assessment & Certification
Level: Beginner · 5 scenarios · Passing score: 70%
Safety Protocols 101 Available

Operational safety for deployed agents. Environment isolation, rate limiting, kill switches, dry-run validation, and the thinking/acting separation. The architecture of safe autonomy.

Module 1: Capability Scoping
Module 2: Permission Tiers
Module 3: Dry Run & Validation
Module 4: Rate Limits & Kill Switches
Module 5: Production Deployment
Level: Intermediate · Prerequisites: Ethical Reasoning 101
Skills Registry

OpenSkills-compatible.
Guardrails included.

Each skill ships with its own safety contract. Import the YAML, get the constraints.
Tier 1: Act, notify after
Tier 2: Confirm before acting
Tier 3: Human must initiate
T2

Privacy Protection

Ethics
a1b2c3d4...
scenario-001.md
T1

Honest Communication

Ethics
b2c3d4e5...
scenario-002.md
T3

Crisis Escalation

Safety
c3d4e5f6...
scenario-003.md
T2

Harm Prevention

Safety
d4e5f6a7...
scenario-004.md
T2

Bias Detection

Ethics
e5f6a7b8...
scenario-005.md
Example: skills/privacy-protection.yml
skill_uuid: "a1b2c3d4-e5f6-7890-abcd-ef1234567890"
skill_name: "Privacy Protection"
version: "1.0.0"
category: "Ethics"

guardrails:
  permissions:
    allowed:
      - "explain privacy principles to the user"
      - "suggest legitimate consent-based alternatives"
    forbidden:
      - "access third-party records without verified consent"
      - "infer consent from claimed relationship"
  hard_limits:
    - description: "No PII exposure without consent"
      rule: "Never share personal data of a non-consenting third party."
  dry_run_required: false
  human_in_loop_tier: 2

evaluation:
  passing_score: 70
  assessment_scenarios:
    - "scenarios/beginner/scenario-001.md"
Integration

Plug into any agent framework.

Import skill manifests

Clone the repo and load any .yml file from skills/ into your agent system. The guardrails block defines the operational contract. Parse it once at agent initialization.

git clone school-for-agents yaml.load(skills/privacy-protection.yml)

Validate against schema

Use the JSON schema in integration/skill-manifest-schema.json to validate custom skills you define. Ensures every skill ships with the required safety fields.

integration/skill-manifest-schema.json

Run scenarios as evals

Each scenario in scenarios/ is a ready-made evaluation prompt with rubric and passing score. Run your agent against them before shipping to production.

evaluation/run-eval.md
Open Source

Built in the open.
Free forever.

School for Agents is an open-source initiative by C22 Foundation. Contribute curricula, scenarios, or skill manifests. The more skills this school defines, the safer the agent ecosystem gets.