Skip to content

COE: CO for Education

Application COE Status In Analysis (pilot planned) License CC BY 4.0 Base CO v1.1

COE is the instructor-side CO domain application. It does not ban AI. It redesigns assessment so that AI collaboration makes student judgment more visible, not less.

The core insight: if students will use AI regardless, design assessments where the interesting signal is how they use it. The COE Spectrum gives instructors four levels of progressive freedom, each shifting what gets assessed.

Four levels of assessment, from full instructor control to full student autonomy:

LevelNameInstructor definesStudent definesWhat gets assessed
1Full ConstraintAll five system components (agents, knowledge, guardrails, workflow, learning)Only prompts and decisionsQuality of thinking within constraints
2Partial FreedomL3 guardrails, L4 workflowL1 agents, L2 knowledge baseAgent design + knowledge architecture + output
3Full FreedomMinimal safety constraints onlyEverythingFull CO setup + output + deliberation journal
4MetaNothingSetup designed for someone else to useSetup usability by others (knowledge transfer)
Level 1Level 2Level 3Level 4
Following instructionsDesigning experimentsDesigning methodologiesTeaching

Each level increases the student’s responsibility for designing the AI collaboration setup. At Level 4, the student creates a setup that another student can use. This is the hardest assessment: it requires genuine understanding of both the domain and the methodology.

COE deploys two categories of agents:

Student-facing (configured by the instructor, used by the student):

AgentRole
Domain ExpertSubject knowledge calibrated to the course level
Methodology GuideWalks students through the assignment workflow

Assessment (instructor-side, evaluates student work):

AgentRole
Rubric ScorerScores each rubric dimension against calibrated standards
Pattern DetectorIdentifies collaboration patterns across student submissions
Anomaly FlaggerDetects statistical anomalies in submission characteristics
Quality AssessorEvaluates overall work quality against course benchmarks
Summary GeneratorProduces assessment summaries for instructor review

COE addresses the “students will just ask AI to do everything” problem at the architecture level:

Assessment targets shift. At Level 1, the output matters. At Level 3, the deliberation journal matters more than the output. At Level 4, the usability of the student’s setup is the assessment. Asking AI to “do everything” produces a poor deliberation journal and an unusable setup.

Audit trails exist. Every student-AI interaction produces a log. The Pattern Detector identifies when students bypass the intended workflow. The Anomaly Flagger catches statistical outliers (e.g., submissions completed in unusually short times, or with unusually uniform quality across rubric dimensions).

The rubric is transparent. Students see the rubric before they begin. This is not a trap. It is a signal: here is what matters. If what matters is judgment quality, students must demonstrate judgment. AI cannot do this for them.

COE defines two parallel workflows:

Student workflow (5 phases):

  1. Setup: receive the assignment, understand the rubric
  2. Exploration: research the topic with AI assistance
  3. Development: produce the deliverable
  4. Refinement: revise based on self-assessment
  5. Submission: package and submit

Instructor workflow (4 phases):

  1. Calibration: define rubric, set scoring standards, seed anti-gaming config
  2. AI Scoring: assessment agents score all submissions
  3. Review: instructor reviews AI scores, adjusts where judgment disagrees
  4. Moderation: inter-rater agreement check (threshold: >0.7 between calibration and scoring)

Three approval gates require human judgment before grades finalise.

Every student action is traceable. COE uses cryptographic audit trails (based on the EATP trust protocol) so instructors can verify the full chain from submission back through the AI collaboration to the student’s original prompts and decisions.

  • Every student action traces to the student (not to AI)
  • Rules define what AI may do at each COE Spectrum level
  • Tamper-evident records document the entire collaboration
  • The instructor can verify who decided what, and when

COE and COL are complementary:

  • COE is the Trust Plane: the instructor designs constraints, rubrics, and assessment criteria
  • COL is the Execution Plane: the student works within those constraints to produce deliverables

This mirrors the Dual Plane Model from CARE.

COE is in analysis with a pilot planned. The links below give access to the current research implementation.

  1. Download the COE workspace from GitHub and unzip it
  2. Open Claude Desktop, switch to the Cowork tab, and open the unzipped folder
  3. Type /start and describe your course and assessment goals

The full setup guide walks through each step.