AI Setup

AI Setup

Choose which prompt variant each agent uses. Activation takes effect on the next agent run.

AI Use Case Analyst

Active: H
A — full context (baseline)
Original prompt: full use-case database, full production JSON, full extracted source-document text, full previous-proposal payload. Highest cost, used as comparison baseline.
generator
reviewer
B — compact context
Slim use-case fields, stripped production JSON (drops the role catalogue and audit fields), summary-only for long source documents, and compact previous-proposal bullets. ~60% fewer input tokens vs. Variant A.
generator
reviewer
C — markdown context
Production rendered as markdown — overview table, description prose, then workflow grouped by stage and subsection with each step's fields in a table. Aims to reduce process-step-ID hallucination by removing the JSON nesting and giving each ID its own code-styled table cell.
generator
reviewer
D — tool-based discovery
No use-case catalogue in the prompt. The model discovers use cases on demand by calling a `search_use_cases` tool (exact filters, text search, vector search). Defaults to excluding use cases already linked into this production's workflow.
generator
reviewer
E — forbidden pairs (markdown)
Like Variant C but the catalogue's `Already Mapped` column is removed and replaced with a top-of-prompt 'Forbidden Pairs' block that explicitly lists every (useCaseId, processStepId) currently on the production. Reduces the model's tendency to misread the column as a recommendation about which step a use case should target.
generator
reviewer
F — production summary (markdown)
Like Variant E, but raw source-document text is replaced by a cached production-level summary tailored for AI use case identification. The summary is regenerated only when the active source-document set changes. Falls back to Variant E's per-doc summaries if summary generation fails.
generator
reviewer
G — review-focused (catalogue + fit-check)
Generator behaviour matches Variant F (cached document summary, forbidden pairs, full catalogue). The reviewer pass is refocused: dead validation summaries and the verbose proposal-history dump are removed; the use case catalogue is added so the reviewer can flag pairs where a clearly better use case exists for the proposed step. Scoring rubric narrows to step fit, application specificity, and better-fit-available.
generator
reviewer
H — code reviews only (no AI review)
Active
Generator behaviour matches Variant F with two constraints layered on: the system prompt caps output at 50 (useCase, processStep) mappings, and max_tokens is reduced to 50 000. The AI reviewer pass is skipped entirely; only the deterministic code-level checks (UC name validation, process step ID validation, feedback compliance, already-mapped pair stripping) gate the proposal. Opportunities that pass the hard checks are auto-approved. Useful as a baseline for measuring the AI review's contribution.
generator
Used for the next run.

Role Assignment Analyst

Active: D
A — full context (baseline)
Original prompt: full role catalog JSON in the system prompt, full production JSON, raw extracted source-document text, full known-members section, full JSON dump of the 3 most recent completed proposals. Highest cost, used as comparison baseline.
generator
reviewer
B — production summary (markdown)
Compact markdown role catalogue plus markdown-rendered production. Raw source-document text is replaced by a cached production-level summary tailored for role assignment (named people, departments, vendors, scale signals). The summary is regenerated only when the active source-document set changes. Falls back to raw source-document text if summary generation fails.
generator
reviewer
C — AI-driven carry-forward (no rationale)
Like Variant B, but the prompt no longer instructs the model to produce a `previousItemsRationale` array. Pending and rejected roles are listed informationally; the AI decides which (if any) to repropose. Novelty is enforced by code: any role already accepted on the production is dropped before review.
generator
reviewer
D — proposal-step format (no previous proposals)
Active
Like Variant C, but the production is rendered without owner/contributor columns (matches the workflow-steps proposal-step format) and the compact previous-proposals section is omitted. Pending and rejected roles remain in the prompt; novelty is enforced in code.
generator
reviewer
Used for the next run.

Workflow Steps Architect

Active: E
A — full context (baseline)
Original prompt: full production document, raw extracted source-document text, JSON dump of the 3 most recent completed proposals, full action-rule restatement in the user prompt. Highest cost, used as comparison baseline.
generator
reviewer
B — production summary (markdown)
Production rendered as markdown (overview + per-stage workflow tables); previous proposals as compact bullets, accepted steps filtered out; raw source-document text replaced by a cached production-level summary tailored for workflow planning. Tools narrowed to ontology search (get_organization dropped — org info inlined). Current Team section removed (Role agent's domain). AI review runs against the same production markdown, org info, source-document summary, and previous proposals the generator saw. Falls back to raw source-document text if summary generation fails.
generator
reviewer
C — two-pass generator (no review)
First pass uses Variant B's prompt and tool set to draft a proposal. A second generator pass receives the draft and is asked to self-critique and return a refined proposal — same JSON schema, no extra context. AI review is skipped (the refinement pass is the review). Deterministic code checks still gate the final proposal.
generator
refinement
D — global library reference (no review)
VariantB's prompt and tools, plus a Global Workflow Steps Library section listing entries from past productions of the same type, grouped by stage and subsection and sorted by usage count. The AI reviewer pass is skipped — only deterministic code checks gate the proposal. The library section is omitted when no entries match the production's type.
generator
E — global library, no descriptions, no review
Active
VariantD's prompt skeleton (production markdown, source-doc summary, Global Workflow Steps Library, ontology-only tools, AI review skipped) but the model's output schema has NO `description` field on steps. The step `name` carries its meaning; `actionRationale` carries the why. `modify` actions are assets-only — a modify is valid only when the new `assets` list differs from the existing step.
generator
Used for the next run.
F — E + production notes summary
VariantE's prompt skeleton (production markdown, source-doc summary, Global Workflow Steps Library, ontology-only tools, AI review skipped, no `description` field on output steps, modify-assets-only) plus a cached summary of the production's communications/notes between source docs and library. The notes summary weights more-recent notes higher when they conflict with older ones. Cache regenerates on note adds, deletes, or edits.
generator