-
Notifications
You must be signed in to change notification settings - Fork 0
docs(research): update agent ecosystem with OpenAI Agents SDK insights #22189
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -75,3 +75,4 @@ tsconfig.json | |
| **/tsconfig.json | ||
| tsconfig.*.json | ||
| **/tsconfig.*.json | ||
| node_modules_corrupted | ||
| Original file line number | Diff line number | Diff line change | ||||
|---|---|---|---|---|---|---|
|
|
@@ -2,11 +2,23 @@ | |||||
|
|
||||||
| ## Executive Summary | ||||||
|
|
||||||
| The AI agent ecosystem has matured significantly, moving past experimental phases into production-grade orchestration systems. Three prominent frameworks continue to dominate the multi-agent landscape in 2026: **LangGraph**, **CrewAI**, and **AutoGen**. Each addresses different operational paradigms, ranging from strict graph-based state machines to dynamic conversational workflows. | ||||||
| The AI agent ecosystem has matured significantly, moving past experimental phases into production-grade orchestration systems. Prominent frameworks continue to dominate the multi-agent landscape in 2026: **LangGraph**, **CrewAI**, **AutoGen**, and the emerging **OpenAI Agents SDK**. Each addresses different operational paradigms, ranging from strict graph-based state machines to dynamic conversational workflows. There is a notable industry trend towards the "Agent as a Tool" and handoff patterns, offering more modular, transparent, and auditable multi-agent collaboration. | ||||||
|
|
||||||
| ## Framework Analysis & Capabilities | ||||||
|
|
||||||
| ### 1. LangGraph (LangChain) | ||||||
| ### 1. OpenAI Agents SDK (Swarm Evolution) | ||||||
|
|
||||||
| The OpenAI Agents SDK represents a streamlined, native approach to multi-agent orchestration, heavily relying on the concepts of routines and handoffs without needing complex external framework dependencies. | ||||||
|
|
||||||
| - **Core Paradigm:** Agent-as-a-Tool and Handoffs. | ||||||
| - **Key Capabilities:** | ||||||
| - **Native Integration:** Direct integration with OpenAI's APIs, leveraging the newest model capabilities seamlessly. | ||||||
| - **Tool Support:** Comprehensive support for custom Python functions, managed tools (e.g., Code Interpreter, WebSearch), and external MCP servers. | ||||||
| - **Handoff Mechanism:** Agents can seamlessly transfer control to specialized peer agents based on task requirements, treating other agents essentially as executable tools. | ||||||
| - **Strict LLM Orchestration:** Avoids heavy state-machine abstractions in favor of letting the LLM's tool-calling logic drive the orchestration flow directly. | ||||||
| - **Best Use Cases:** Systems requiring transparent, auditable collaboration with minimal orchestration boilerplate, leveraging specialized sub-agents. | ||||||
|
|
||||||
| ### 2. LangGraph (LangChain) | ||||||
|
|
||||||
| LangGraph has solidified its position as the premier framework for complex, stateful, and deterministic orchestration. With the stable release of LangChain 1.0 and LangGraph 1.0, it excels in environments with strict auditability and high-reliability requirements. | ||||||
|
|
||||||
|
|
@@ -19,7 +31,7 @@ LangGraph has solidified its position as the premier framework for complex, stat | |||||
| - **Stability and Modernization:** Python 3.10+ requirement and simplified package structure for production-grade deployments. | ||||||
| - **Best Use Cases:** Complex, conditional pipelines; production systems requiring compliance and strict audit trails. | ||||||
|
|
||||||
| ### 2. CrewAI | ||||||
| ### 3. CrewAI | ||||||
|
|
||||||
| CrewAI focuses on simplifying the creation of multi-agent systems by leveraging intuitive human-like team metaphors. It offers the fastest path from prototype to functional multi-agent collaboration. | ||||||
|
|
||||||
|
|
@@ -31,7 +43,7 @@ CrewAI focuses on simplifying the creation of multi-agent systems by leveraging | |||||
| - **MCP Integration:** Native support for the Model Context Protocol (MCP), enabling deeper integration with external tools and resources. | ||||||
| - **Best Use Cases:** Business workflows, research syndication, and task delegation where roles map neatly to human organizational structures. | ||||||
|
|
||||||
| ### 3. AutoGen (Microsoft Agent Framework) | ||||||
| ### 4. AutoGen (Microsoft Agent Framework) | ||||||
|
|
||||||
| Backed by enterprise resources and now in version 0.4.0+, AutoGen excels in dynamic, conversational interactions and complex problem-solving where iterative refinement is required. | ||||||
|
|
||||||
|
|
@@ -45,9 +57,10 @@ Backed by enterprise resources and now in version 0.4.0+, AutoGen excels in dyna | |||||
|
|
||||||
| ## Industry Trends & Next Steps | ||||||
|
|
||||||
| - **Hybrid Architectures:** We are seeing an increase in production deployments combining frameworks (e.g., LangGraph for overall state orchestration, wrapping a CrewAI team for a specific research sub-task). | ||||||
| - **Hybrid Architectures:** We are seeing an increase in production deployments combining frameworks (e.g., LangGraph for overall state orchestration, wrapping a CrewAI team or an OpenAI Agent SDK routine for a specific research sub-task). | ||||||
|
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Use one canonical name for the framework across the doc.
✏️ Proposed wording fix-- **Hybrid Architectures:** We are seeing an increase in production deployments combining frameworks (e.g., LangGraph for overall state orchestration, wrapping a CrewAI team or an OpenAI Agent SDK routine for a specific research sub-task).
+- **Hybrid Architectures:** We are seeing an increase in production deployments combining frameworks (e.g., LangGraph for overall state orchestration, wrapping a CrewAI team or an OpenAI Agents SDK routine for a specific research sub-task).
-_Update:_ We have explicitly expanded our benchmarks to track State Recovery Success Rate (SRSR), Coordination Token Overhead (CTO), and Orchestration Latency Penalty (OLP). We have also created adapter layers for LangGraph, CrewAI, AutoGen, and OpenAI Agents to support these metrics.
+_Update:_ We have explicitly expanded our benchmarks to track State Recovery Success Rate (SRSR), Coordination Token Overhead (CTO), and Orchestration Latency Penalty (OLP). We have also created adapter layers for LangGraph, CrewAI, AutoGen, and OpenAI Agents SDK to support these metrics.Also applies to: 66-66 🤖 Prompt for AI Agents |
||||||
| - **Agent as a Tool:** A massive shift towards the "Agent as a Tool" handoff pattern (popularized by OpenAI Agents SDK) where central orchestrators treat specialized sub-agents simply as functional tool calls. | ||||||
| - **Production Safety:** Error handling and robust fallback mechanisms ("safe nodes") are becoming standard requirements over sheer capability. | ||||||
|
|
||||||
| **Recommendation:** Summit's internal orchestration and benchmarking must expand to cover these advanced topologies, specifically evaluating the overhead of coordination and the resilience of durable execution under load. | ||||||
|
|
||||||
| _Update:_ We have explicitly expanded our benchmarks to track State Recovery Success Rate (SRSR), Coordination Token Overhead (CTO), and Orchestration Latency Penalty (OLP). We have also created adapter layers for LangGraph, CrewAI, and AutoGen to support these metrics. | ||||||
| _Update:_ We have explicitly expanded our benchmarks to track State Recovery Success Rate (SRSR), Coordination Token Overhead (CTO), and Orchestration Latency Penalty (OLP). We have also created adapter layers for LangGraph, CrewAI, AutoGen, and OpenAI Agents to support these metrics. | ||||||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. For consistency with other references in this document (like the new section heading) and the related
Suggested change
|
||||||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
🧩 Analysis chain
🌐 Web query:
Does actions/setup-node v4 install pnpm automatically?💡 Result:
No, actions/setup-node v4 does not install pnpm automatically. It sets up Node.js and supports caching for pnpm (when cache: 'pnpm' is specified and pnpm v6.10+), but pnpm must be pre-installed or installed separately (e.g., via pnpm/action-setup or corepack). The official documentation explicitly states "Package manager should be pre-installed" for the cache input, and GitHub Actions runners do not have pnpm pre-installed (unlike yarn).
Citations:
🏁 Script executed:
cat -n .github/workflows/schema-change-check.yml | head -120Repository: BrianCLong/summit
Length of output: 4255
🏁 Script executed:
Repository: BrianCLong/summit
Length of output: 235
🏁 Script executed:
Repository: BrianCLong/summit
Length of output: 959
🏁 Script executed:
Repository: BrianCLong/summit
Length of output: 2300
🏁 Script executed:
Repository: BrianCLong/summit
Length of output: 154
pnpm is not installed before use — workflow will fail.
The step uses
pnpm installbut pnpm is not available by default on GitHub-hosted runners. Node.js 24 ships with corepack, but it's not enabled on GitHub Actions runners. This will result inpnpm: command not found.Add a step to enable corepack before using pnpm:
Alternatively, use the repository's existing
setup-pnpmcustom action (.github/actions/setup-pnpm) which handles this automatically.🤖 Prompt for AI Agents