There are a handful of well-known agent frameworks and platforms, and you know what? They can all demonstrate the value of agents with a simple proof of concept. They all support agent planning, orchestration, memory management and tool use – the building blocks of agentic workflows. And yes, they’re all opinionated frameworks. However, the differences in how they’ve implemented these building blocks are not what matters most.
It’s one thing to support a proof of concept. It’s another to support 100s, let alone 1000s, of agentic workflows running in production. And this is what matters, the ability to grow from performing a simple task via agent collaboration to effectively supporting every role and department via autonomous agents handling complex, critical tasks – or in CrewAI language, becoming agent-native.
MCP (model context protocol)
MCP has successfully standardized how agents call tools and access data sources, with 1,000s of servers available for everything from cloud services and developer tools to file systems and security platforms. With support for MCP, it’s easier than ever to build powerful agentic workflows that determine which tools to access and when.
Build agents that can:
serverparams = {"url": "http://localhost:8000/sse"}
with MCPServerAdapter(serverparams) as tools:
# tools is now a list of CrewAI Tools matching 1:1 with the MCP server's tools
agent = Agent(..., tools=tools)
task = Task(...)
crew = Crew(..., agents=[agent], tasks=[task])
crew.kickoff(...)
That’s not all.
CrewAI Enterprise now includes bidirectional MCP support, meaning crews and flows can be accessed by remote MC clients, whether desktop applications, web applications and servers or even remote agents – allowing customers to fully integrate CrewAI Enterprise with existing infrastructure and systems, and become fully agent native.
Tools repository
As enterprises look to expand their use agents and become agent native, a centralized means of deploying and managing custom tools which can be accessed by different agents in different workflows will be necessary to scale.
CrewAI Enterprise now supports private tool repositories for customers, and in keeping with high security standards, access to tools is managed via role-based access control. Whereas public tools are available to anyone in an organization building agents, private tools require explicit permission – as does the ability to publish tools to the repository.

Guardrails
AI assistants and copilots shouldn’t be deployed to production without guardrails, and neither should agents. They’re often overlooked when experimenting with agents and building simple prototypes, but they’re necessary to ensure the automation of critical business processes via agents can be relied upon – both in terms of correctness as well as behaviour.
CrewAI allows developers to add guardrails via functions or LLM-as-a–Judge prompts. For example, a function to check the output length (e.g., too long) or see if the output contains specific keywords or phrases (e.g., competitor mentions). And for more subjective checks, you can specify a prompt, and optionally, the LLM (e.g., Llama Guard 3).
blog_task = Task(
description="Write a blog post about AI",
expected_output="A blog post under 200 words",
agent=blog_agent,
guardrail=TaskGuardrail(
description="Ensure the blog is less than 200 words",
llm=LLM(model="gpt-4o-mini"),
)
)
CrewAI Enterprise provides customers with an advanced, out-of-the-box guardrail for hallucination detection as well.
Events and webhooks
A powerful eventing system is required for any infrastructure platform, and provides developers with much needed support for extending it with everything from custom debugging and monitoring to advanced integration with other platforms and services. In addition, events and webhooks can be used to add evaluation capabilities or to integrate with external evaluation platforms.
CrewAI now includes an event bus, allowing developers to create listeners which run when any of the following events are emitted:
CrewAI Enterprise customers can register webhooks with authentication tokens for any of these events, making it even easier to integrate with other platforms and services.
{
"inputs": { "foo": "bar"},
"taskWebhookUrl": "", // unchanged for compatibility
"stepWebhookUrl": "", // unchanged for compatibility
"crewWebhookUrl": "", // unchanged for compatibility
"webhooks": {
"events": ["crew_kickoff_started", "llm_call_started"], // list of events to subscribe to
"url": "https://webhook.test/foo/bar", // url to send events to
"authentication": {
"strategy": "bearer",
"token": "de82dc64-8e4c-4270-81fb-ac9668a79f5e" // sent in every webhook request over headers
}
}
}
Agentic RAG
With the introduction of retrieval augmented generation (RAG) in 2020 by researchers at Meta, external knowledge has played a critical role – especially within enterprise AI applications. However, RAG is now giving way to agentic RAG, with agents determining which knowledge base to access and how.
CrewAI has added a number of features to not support agentic RAG, but to provide developers with greater flexibility.
# Original task prompt
task_prompt = "Answer the following questions about the user's favorite movies: What movie did John watch last week? Format your answer in JSON."
# Behind the scenes, this might be rewritten as:
rewritten_query = "What movies did John watch last week?"
The time for Agentic AI is now
CrewAI is the most powerful Agentic AI platform available, but more importantly, it’s a platform enterprises can rely on and trust to power their AI transformation. The team at CrewAI is driven by a passionate desire to unlock the full potential of Agentic AI. However, it requires much more than building new APIs. It requires a commitment to delivering the features and capabilities organizations need to succeed with it.
CrewAI's first launch week has gotten off to a great start, but we have a lot more to share.
Sign up for our webinar this Thursday, and discover what’s possible with Agentic AI today!
Manage the full AI agent lifecycle — build, test, deploy, and scale — with a visual editor and ready-to-use tools.
All the power of AMP Cloud, deployed securely on your own infrastructure — on-prem or private VPCs in AWS, Azure, or GCP
An open-source orchestration framework with high-level abstractions and low-level APIs for building complex, agent-driven workflows.