Start Agent monitoring, evaluation, and observability
Maxim AI provides comprehensive agent monitoring, evaluation, and observability for your CrewAI applications. With Maxim’s one-line integration, you can easily trace and analyse agent interactions, performance metrics, and more.
Maxim’s Prompt Management capabilities enable you to create, organize, and optimize prompts for your CrewAI agents. Rather than hardcoding instructions, leverage Maxim’s SDK to dynamically retrieve and apply version-controlled prompts.
Create, refine, experiment and deploy your prompts via the playground. Organize of your prompts using folders and versions, experimenting with the real world cases by linking tools and context, and deploying based on custom logic.
Easily experiment across models by configuring models and selecting the relevant model from the dropdown at the top of the prompt playground.
Create, refine, experiment and deploy your prompts via the playground. Organize of your prompts using folders and versions, experimenting with the real world cases by linking tools and context, and deploying based on custom logic.
Easily experiment across models by configuring models and selecting the relevant model from the dropdown at the top of the prompt playground.
As teams build their AI applications, a big part of experimentation is iterating on the prompt structure. In order to collaborate effectively and organize your changes clearly, Maxim allows prompt versioning and comparison runs across versions.
Iterating on Prompts as you evolve your AI application would need experiments across models, prompt structures, etc. In order to compare versions and make informed decisions about changes, the comparison playground allows a side by side view of results.
Prompt comparison combines multiple single Prompts into one view, enabling a streamlined approach for various workflows:
Maxim AI provides comprehensive observability & evaluation for your CrewAI agents, helping you understand exactly what’s happening during each execution.
Track your agent’s complete lifecycle, including tool calls, agent trajectories, and decision flows effortlessly.
Track your agent’s complete lifecycle, including tool calls, agent trajectories, and decision flows effortlessly.
Run detailed evaluations on full traces or individual nodes with support for:
Evaluate captured logs automatically from the UI based on filters and sampling
Use human evaluation or rating to assess the quality of your logs and evaluate them.
Evaluate any component of your trace or log to gain insights into your agent’s behavior.
Set thresholds on error, cost, token usage, user feedback, latency and get real-time alerts via Slack or PagerDuty.
Visualize Traces over time, usage metrics, latency & error rates with ease.
Install the Maxim SDK via pip:
Or add it to your requirements.txt
:
That’s it! All your CrewAI agent interactions will now be logged and available in your Maxim dashboard.
Check this Google Colab Notebook for a quick reference - Notebook
After running your CrewAI application:
Log in to your Maxim Dashboard
Navigate to your repository
View detailed agent traces, including:
No traces appearing: Ensure your API key and repository ID are correct
Ensure you’ve called instrument_crewai()
before running your crew. This initializes logging hooks correctly.
Set debug=True
in your instrument_crewai()
call to surface any internal errors:
Configure your agents with verbose=True
to capture detailed logs:
Double-check that instrument_crewai()
is called before creating or executing agents. This might be obvious, but it’s a common oversight.
Official CrewAI documentation
Official Maxim documentation
Maxim Github