Maxim Integration
Start Agent monitoring, evaluation, and observability
Maxim Overview
Maxim AI provides comprehensive agent monitoring, evaluation, and observability for your CrewAI applications. With Maxim’s one-line integration, you can easily trace and analyse agent interactions, performance metrics, and more.
Features
Prompt Management
Maxim’s Prompt Management capabilities enable you to create, organize, and optimize prompts for your CrewAI agents. Rather than hardcoding instructions, leverage Maxim’s SDK to dynamically retrieve and apply version-controlled prompts.
Create, refine, experiment and deploy your prompts via the playground. Organize of your prompts using folders and versions, experimenting with the real world cases by linking tools and context, and deploying based on custom logic.
Easily experiment across models by configuring models and selecting the relevant model from the dropdown at the top of the prompt playground.
Create, refine, experiment and deploy your prompts via the playground. Organize of your prompts using folders and versions, experimenting with the real world cases by linking tools and context, and deploying based on custom logic.
Easily experiment across models by configuring models and selecting the relevant model from the dropdown at the top of the prompt playground.
As teams build their AI applications, a big part of experimentation is iterating on the prompt structure. In order to collaborate effectively and organize your changes clearly, Maxim allows prompt versioning and comparison runs across versions.
Iterating on Prompts as you evolve your AI application would need experiments across models, prompt structures, etc. In order to compare versions and make informed decisions about changes, the comparison playground allows a side by side view of results.
Why use Prompt comparison?
Prompt comparison combines multiple single Prompts into one view, enabling a streamlined approach for various workflows:
- Model comparison: Evaluate the performance of different models on the same Prompt.
- Prompt optimization: Compare different versions of a Prompt to identify the most effective formulation.
- Cross-Model consistency: Ensure consistent outputs across various models for the same Prompt.
- Performance benchmarking: Analyze metrics like latency, cost, and token count across different models and Prompts.
Observability & Evals
Maxim AI provides comprehensive observability & evaluation for your CrewAI agents, helping you understand exactly what’s happening during each execution.
Track your agent’s complete lifecycle, including tool calls, agent trajectories, and decision flows effortlessly.
Track your agent’s complete lifecycle, including tool calls, agent trajectories, and decision flows effortlessly.
Run detailed evaluations on full traces or individual nodes with support for:
- Multi-step interactions and granular trace analysis
- Session Level Evaluations
- Simulations for real-world testing
Auto Evals on Logs
Evaluate captured logs automatically from the UI based on filters and sampling
Human Evals on Logs
Use human evaluation or rating to assess the quality of your logs and evaluate them.
Node Level Evals
Evaluate any component of your trace or log to gain insights into your agent’s behavior.
Set thresholds on error, cost, token usage, user feedback, latency and get real-time alerts via Slack or PagerDuty.
Visualize Traces over time, usage metrics, latency & error rates with ease.
Getting Started
Prerequisites
- Python version >=3.10
- A Maxim account (sign up here)
- Generate Maxim API Key
- A CrewAI project
Installation
Install the Maxim SDK via pip:
Or add it to your requirements.txt
:
Basic Setup
1. Set up environment variables
2. Import the required packages
3. Initialise Maxim with your API key
4. Create and run your CrewAI application as usual
That’s it! All your CrewAI agent interactions will now be logged and available in your Maxim dashboard.
Check this Google Colab Notebook for a quick reference - Notebook
Viewing Your Traces
After running your CrewAI application:
-
Log in to your Maxim Dashboard
-
Navigate to your repository
-
View detailed agent traces, including:
- Agent conversations
- Tool usage patterns
- Performance metrics
- Cost analytics
Troubleshooting
Common Issues
-
No traces appearing: Ensure your API key and repository ID are correct
-
Ensure you’ve
called instrument_crewai()
before running your crew. This initializes logging hooks correctly. -
Set
debug=True
in yourinstrument_crewai()
call to surface any internal errors: -
Configure your agents with
verbose=True
to capture detailed logs: -
Double-check that
instrument_crewai()
is called before creating or executing agents. This might be obvious, but it’s a common oversight.