Arti Agent Stack is built on the principles of Compositional Agentic Architecture (CAA) β a modular, observable, collaboration-first approach to intelligent system design.
This repo introduces the CAA philosophy, outlines its 5-layer architecture, and defines the 10 operational principles behind Arti's implementation.
Chatbots answer questions. Enterprises need work done.
Most agent frameworks are brittle, opaque, and stuck in the demo phase.
CAA turns agents from toys into systems.
CAA defines five layers that turn LLM reasoning into reliable execution.
CAA is a blueprint, not a framework.
| Layer | Solves | Delivers |
|---|---|---|
| Context | AI doesnβt get intent | Typed, versioned context |
| Behavior | Reasoning is hidden | Explicit, inspectable plans |
| Execution | Chats but no action | Deterministic workflows & tool calls |
| State | AI forgets | Persistent, testable memory |
| Collaboration | No trust / approvals | Human-in-loop hooks |
| Observability (cross-cutting) | Opaque failures | Full tracing & metrics |
| Security (cross-cutting) | Risk & compliance gaps | Auth, isolation, governance |
| Learning & Adaptation (lifecycle) | Systems decay | Continuous improvement loop |
If you're building agents that need to survive contact with reality, start here. Weβve tested dozens of frameworks, shipped agent systems in production, and learned what works. This project distills that into a blueprint for real-world AI execution β not just chatbots and demos.
Overview
Diagram
%% Compositional Agentic Architecture β data-flow view
flowchart LR
%% Core loop
Context["Context<br>Layer"] --> Behavior["Behavior<br>Layer"]
Behavior --> Execution["Execution<br>Layer"]
Execution --> State["State<br>Layer"]
State -->|state_snapshot| Context
%% Human collaboration (side-channel)
Execution -. tool_events .-> Collaboration["Collaboration<br>Layer"]
State -. snapshot .-> Collaboration
Collaboration -. feedback .-> Behavior
Collaboration -. override .-> Execution
%% Observability bus (cross-cutting)
Context -. ctx_trace .-> Obs["Observability<br>Bus"]
Behavior -. plan_trace .-> Obs
Execution -. act_trace .-> Obs
State -. state_trace .-> Obs
Collaboration -. collab_trace .-> Obs
Already have a LangChain, Pydantic, or OpenAI codebase? You don't need to rewrite everything to get the benefits of Compositional Agentic Architecture.
CAA is a set of architectural patterns, not a rigid framework. π Read the Integration Guide to see how to organize your existing tools (Pydantic, Redis, Celery) into robust agentic layers.
CAA replaces "one big prompt" with specialized layers:
# Instead of: response = llm.chat(user_input)
# 1. Context Layer (Validate)
context = UserContext.from_input(user_input) # Enforces Pydantic Schema
# 2. Behavior Layer (Plan)
plan = planner.create_plan(context) # Deterministic steps, no execution yet
# 3. Execution Layer (Act)
result = executor.run(plan) # Safe tool use with contracts
# 4. State Layer (Remember)
state.save(context, plan, result) # Persist to Redis- For Leaders & Product Managers: Start with our Product Overview to understand the business value.
- For Architects & Engineers: Dive into the CAA Architectural Stack for the technical deep dive.
- For the Curious Reader: Check the Philosophy to understand the motivation & background.
- Execution-first AI β Not chatbots. Not search. Agents that do.
- Typed prompts and context β No raw string stitching.
- Composable behaviors β One agent, one responsibility.
- Separation of concerns β Prompting β tooling β state.
- Observability by design β Trace every decision and step.
CAA is not just a list of ideas; it's a structured, 5-layer stack designed for production realities. Each layer has a distinct responsibility, ensuring a clear separation of concerns.
- Layer 1: Context - Transforms raw data into a structured understanding, so the AI acts with full context, not guesswork.
- Layer 2: Behavior - The "planner" that decides what to do next, creating an explicit, inspectable execution plan.
- Layer 3: Execution - The "engine room" that reliably performs the actions and calls the tools defined in the plan.
- Layer 4: State** - The persistent memory that tracks the agent's progress and enables multi-step, resumable workflows.
- Layer 5: Collaboration** - The interface that puts humans in the loop for approvals, oversight, and safe operations.
Plus, 2 cross-cutting concern that touches every layer:
- The Observability Bus - Provides the full tracing, metrics, and debugging capabilities required to trust the system in production.
- Security - Authentication, isolation, compliance, governance.
and a Lifecycle concern:
- Learning & Adaptation - Continuous improvement from feedback, monitoring, and retraining.
It defines what agentic systems should be: structured, semantic, observable, and designed for human-AI collaboration.
In a world full of fragile demos and prompt loops, CAA sets the architectural foundation for scalable, production-grade AI systems.
The Arti Agent Stack is our opinionated implementation of CAA β forged through real-world deployment in high-stakes environments.
It is one way to implement the blueprint β not the only way.
- Arti provides:
- First-class support for context versioning
- Deterministic execution with retries and validation
- State persistence for long-running workflows
- Human-in-loop hooks for approvals and escalations
- Observability baked in (tracing, metrics, dashboards)
These are the design rules that make the Arti Stack robust, modular, and production-ready:
-
- One agent, one responsibility. Scope tightly.
-
- Divide prompting, tool logic, memory, context, and execution.
- No monoliths. No entangled loops.
-
- Execution should be defined, inspectable, and testable β not inferred.
-
- No string soup. Use typed, versioned context builders.
-
- Prompts are versioned, parameterized, testable, and fallback-safe.
-
- Tools have typed interfaces, clear intent, and predictable side effects.
-
- Every agent step must be traced, debugged, and explainable.
-
- No loose dicts or memory blobs. Use structured, persistent agent state.
-
- Treat failure like a system design concern. Retry, fallback, escalate.
- Support interruption, supervision, review, and override from the start.
- Visual diagrams of the CAA architecture and Arti stack
- Integration Guide for existing tools
- Core Interface Definitions (See /src)
- Example: Industrial Support Agent
The Thinking Behind the Stack
Read our Philosophy to understand the core principles behind CAA and how they evolved from real-world production challenges.
Deep Dive Blog Series
Weβre documenting our journey, lessons, and frameworks in a multi-part blog series. Start with βThe Workforce Challenges in Complex & Technical Industriesβ for the why β and follow the trail to the how.
Follow the Project
Weβre evolving the Arti Agent Stack and Compositional Agentic Architecture in the open.
See It in Action
We're using this stack to power real-world agentic systems in complex industrial environments. Demos and case studies coming soon β follow us on LinkedIn or check out website for updates.
Contribute or Collaborate
Have thoughts, ideas, or real use cases to share? Please follow the guidelines here.
Letβs build the cognitive layer for enterprise AI.
