Skip to content

AI Collaboration Model

The UCCA-AI Rosetta Stone — Human-AI Collaboration Model

Non-binding. Explanatory. Historical. Audience: human architects, maintainers, future collaborators.


Purpose

This document explains how the UCCA Engine emerged through a structured and repeatable human-AI collaboration pattern.

It does not define engine behaviour. It does not impose constraints. It does not function as a contract.

Its purpose is to capture the meta-architecture of collaboration that allowed a general-purpose language model to participate reliably in the design of a deterministic, constraint-driven system.

This document exists to prevent future misinterpretation of how and why the engine took its present form.


1. Actors & Identity

"Alex" — Persona Grounding / Role Conditioning

Defining Alex is an act of persona grounding.

By explicitly constraining the AI to operate as a senior systems engineer and architectural peer, output variance is reduced and inference is biased toward:

  • deterministic reasoning
  • constraint awareness
  • structural consistency
  • long-horizon thinking

This is not anthropomorphism.

It is role-conditioned inference applied deliberately to reduce stochastic behaviour and hallucination risk.

"Tim & Alex" Dynamic — Simulated Multi-Agent Reasoning

Although only a single model is involved, the explicit separation between:

  • Human Architect / Auditor (Tim)
  • AI Logical Executor / Systems Thinker (Alex)

creates a simulated multi-agent environment.

The model must internally reconcile role-specific constraints before producing output. This increases reasoning depth and mirrors multi-agent coordination, even when no explicit chain-of-thought is requested.

The human remains the final authority at all times.


2. The Environment ("The Logic Cage")

Logic Cage — Hard Constraint Satisfaction

The "logic cage" is a set of explicit hard constraints.

Unlike soft stylistic guidance (e.g. "try to keep this clean"), these constraints define invalid solution space.

Outputs that violate constraints are rejected outright.

This approach dramatically narrows the model's search space and increases determinism and repeatability.

Hall of Rejections — Negative Prompting / Search Space Pruning

The "Hall of Rejections" functions as negative prompting.

By explicitly listing prohibited patterns (e.g. unsafe SQL practices, architectural anti-patterns, invalid assumptions), large branches of the probability tree are pruned early.

This preserves attention for valid solution paths and prevents subtle regressions.


3. Data & Memory Management

Handover Briefs — Context State Transfer

The AI does not retain state between sessions.

Handover briefs act as explicit context state transfer artifacts, rehydrating:

  • architectural decisions
  • constraints
  • known trade-offs
  • current system state

This prevents re-litigation of settled truths and protects architectural continuity across sessions.

Memory Index — Externalised Knowledge Base (RAG-lite)

00_READ_FIRST__MEMORY_INDEX.md functions as a manual retrieval-augmented generation system.

Rather than automated embedding search, the human acts as the retrieval layer, ensuring the AI remains grounded in:

  • authoritative documents
  • correct precedence
  • current system reality

This approach avoids accidental hallucination of non-existent structures.

Saturated Window — Context Exhaustion & Attention Drift

As the context window fills, early instructions lose salience and attention drift occurs.

The handover/reset process acts as a context flush, restoring signal-to-noise ratio and reasserting foundational truths.

This is a deliberate operational practice.


4. Process Dynamics

Sequential File Reads — Linear Priming

Enforcing sequential file reads ensures foundational architectural truths are loaded into the model's short-term working memory (KV cache) before complex reasoning begins.

This mirrors compiler design:

  • parse structure first
  • optimise later

Skipping this step increases risk of superficial correctness and deep inconsistency.

Proactive Document Requests — Emergent Task Decomposition

When the AI requests additional documents unprompted, it reflects emergent task decomposition.

The model has inferred that certain sub-tasks (e.g. reading the Memory Index) are required to satisfy the higher-order goal of correctness and consistency.

This behaviour is encouraged and treated as a signal of alignment, not autonomy.


5. Summary Table

UCCA Term Formal AI Concept Purpose
Alex Persona Grounding Reduces randomness and hallucination
Logic Cage Hard Constraints Enforces deterministic output
Hall of Rejections Negative Prompting Prunes invalid logic paths
Handover Brief Context State Transfer Prevents memory loss between sessions
Snapshot Grounded State Aligns output with system reality
Memory Index External Knowledge Base Maintains architectural integrity

Important Clarification

This document does not claim that the AI is:

  • autonomous
  • self-directing
  • intelligent in a human sense
  • capable of independent architectural authority

All outcomes are the result of explicit human intent, clear constraints, and carefully managed context.

The AI functions as a constrained reasoning engine, not a decision-maker.


Closing Note

Most systems document what they built.

This document records how it became possible to build it.

That distinction matters.


End of Document

Version History

Version Date Change Author
1.0 2026-03-11 Migrated from engine/ucca-engine/docs/meta/AI_COLLABORATION_MODEL.md Claude Code