Skip to content

Runtime Orchestration Architecture

This document defines the long-term runtime architecture for a voice-driven, AI-orchestrated ConversionIQ platform.

It does not describe the current implementation. It defines the target runtime model that should guide future product, platform, and documentation decisions.

Use this document when making decisions about:

  • multimodal input handling
  • runtime agent orchestration
  • policy-governed tool execution
  • documentation-driven behavior
  • learning and recommendation loops

Related docs:


The entity stack in layers.md describes what the platform is made of:

  • Workspaces
  • Knowledge Bases
  • Agents
  • Channels
  • Apps

This document describes how the platform should behave at runtime.

These are different views:

  • layers.md = business and system entities
  • runtime-orchestration.md = execution model

They must complement each other and must not be conflated.


ConversionIQ should evolve toward a policy-governed, documentation-driven runtime orchestration system where:

  • voice, text, and UI interactions are normalized into the same intent model
  • runtime behavior is constrained by canonical documentation contracts
  • agents coordinate planning, execution, and validation
  • learning systems improve the platform through governed proposals, not silent mutation of truth

Important clarification:

  • Voice = Text = Prompt is useful at the interaction abstraction level
  • it is not true at the infrastructure level, because voice introduces real-time, privacy, latency, and device-specific concerns that text does not

flowchart TD
inputLayer[InputLayer]
intentLayer[IntentLayer]
policyLayer[PolicyAndGovernanceLayer]
orchestrationLayer[AgentOrchestrationLayer]
executionLayer[ExecutionLayer]
stateLayer[StateAndMemoryLayer]
docsLayer[DocumentationHubAndKnowledgeLayer]
learningLayer[LearningAndOptimizationLayer]
trustLayer[TrustSecurityAndAuditLayer]
inputLayer --> intentLayer
intentLayer --> policyLayer
policyLayer --> orchestrationLayer
orchestrationLayer --> executionLayer
executionLayer --> stateLayer
stateLayer --> learningLayer
docsLayer --> policyLayer
docsLayer --> orchestrationLayer
learningLayer --> docsLayer
trustLayer --> policyLayer
trustLayer --> orchestrationLayer
trustLayer --> executionLayer

Normalizes all inbound signals into a common interaction envelope.

Handles:

  • voice input
  • text input
  • UI interactions
  • channel metadata
  • actor/session/workspace context

Primary output:

  • InteractionEvent

Transforms interaction events into structured, actionable intent.

Handles:

  • intent classification
  • entity extraction
  • ambiguity detection
  • confidence scoring
  • escalation hints

Primary output:

  • StructuredIntent

Determines whether the requested action is allowed, safe, and in scope before orchestration proceeds.

Handles:

  • workspace and tenant isolation
  • RBAC and permission checks
  • compliance rules
  • tool eligibility
  • prompt and workflow selection constraints
  • approval requirements

Coordinates the runtime chain of specialized agents required to satisfy a request.

Handles:

  • agent selection
  • decomposition of tasks
  • context passing
  • handoff sequencing
  • fallback and stop conditions

Carries out the approved action path.

Handles:

  • response generation
  • tool use
  • API calls
  • configuration changes
  • workflow progression
  • state mutation

Maintains the context required for coherent operation and future optimization.

Handles:

  • session state
  • user and workspace context
  • conversation/task memory
  • execution history
  • outcome records

Supplies canonical policy, behavior, and structural knowledge to runtime systems.

Handles:

  • architecture contracts
  • workflow definitions
  • prompt contracts
  • domain rules
  • action and approval classes
  • product behavior references

Observes system usage and turns outcomes into recommendations for improvement.

Handles:

  • pattern detection
  • recommendation generation
  • failure analysis
  • optimization proposals
  • prompt/workflow/doc improvement suggestions

Cross-cutting control layer over the entire runtime.

Handles:

  • audit logging
  • safety and compliance checks
  • prompt injection defenses
  • sensitive action controls
  • policy breach detection
  • human-in-the-loop gating

The future runtime system should use agent roles based on function, not the current delivery workflow roles used by the engineering team.

Recommended runtime agent classes:

  • interprets what the user is trying to achieve
  • identifies missing information
  • determines ambiguity and confidence
  • determines whether the goal is feasible
  • chooses workflow and execution path
  • decomposes work into orchestrated steps
  • validates access, compliance, scope, and action class
  • blocks invalid or unsafe execution paths early
  • performs the actual task
  • invokes tools and services
  • produces user-visible results or mutations
  • checks whether execution satisfied the original goal
  • verifies correctness, safety, and policy alignment
  • captures outcome quality
  • identifies repeated friction or opportunity
  • generates improvement proposals
  • proactively surfaces next-best actions
  • suggests optimizations based on context and prior behavior

The current delivery-agent model in ../ai-governance/decision-flow.md governs how the team builds the platform:

  • Neo
  • Father
  • Coder
  • Helper
  • QA

These are not the same as runtime agents.

Separation rule:

  • docs/ai-governance/** governs engineering and delivery
  • runtime agents belong to product/runtime architecture and should be governed by:
    • docs/architecture/**
    • docs/workflows/**
    • docs/prompts/**
    • future runtime-specific API and tool contracts

This separation is required to avoid mixing:

  • engineering permissions with customer-facing runtime permissions
  • implementation workflows with product behavior
  • local Cursor tooling with runtime orchestration logic

  1. The Documentation Hub is a governed contract system, not a passive wiki.
  2. Runtime systems may read canonical docs and policies directly.
  3. Learning systems should usually produce proposals for documentation or prompt updates, not silently rewrite truth.
  4. Sensitive actions must pass through policy, approval, and audit boundaries.
  5. Workspace and tenant isolation remain non-negotiable runtime constraints.
  6. Runtime orchestration must remain observable, debuggable, and reviewable.
  7. Prefer a small number of reliable runtime agent classes over a large opaque swarm.

This vision is aligned with current documented platform principles:

It extends the current architecture toward:

  • multimodal interaction handling
  • runtime orchestration
  • documentation-driven behavior selection
  • governed learning loops

It does not imply immediate implementation of:

  • full voice infrastructure
  • autonomous self-editing documentation
  • unrestricted tool execution
  • large fully autonomous agent swarms

Focus on contracts and orchestration readiness without requiring full multimodal runtime behavior.

Build now:

  • canonical runtime architecture docs
  • intent contracts for text and UI interactions
  • prompt governance and workflow contracts
  • policy, approval, and action classifications
  • structured telemetry for intent, execution, validation, and outcome
  • proposal-based learning rather than self-editing documentation

Introduce a constrained runtime loop for a small number of high-value tasks.

Build next:

  • planner, executor, and validator pattern
  • recommendation generation
  • runtime policy resolution from approved docs
  • stronger observability and audit traces

Use learning signals to improve the platform through controlled proposals and selective automation.

Build later:

  • workflow improvement proposals
  • prompt revision proposals
  • recommendation tuning
  • governed adaptation within explicit policy boundaries

Phase 4: Multimodal and proactive intelligence

Section titled “Phase 4: Multimodal and proactive intelligence”

Extend the orchestration model to support richer inputs and anticipatory system behavior.

Build much later:

  • voice input and output
  • streaming session management
  • proactive recommendation and optimization flows
  • richer memory and context propagation
  • more advanced specialized runtime agents

  • documentation-driven system contracts
  • intent normalization for text and UI
  • approval-aware orchestration design
  • recommendation and learning telemetry
  • strict separation between canonical truth and runtime proposals
  • direct runtime consumption of fully compiled documentation contracts
  • autonomous multi-step system reconfiguration
  • full voice-first interaction design
  • self-optimizing orchestration loops with limited human intervention