Monitor OpenAI APIs
AppDynamics Python Agent helps you monitor the OpenAI API usage and related metrics. It monitors the ChatCompletion
, Completion
, and Moderation APIs
type requests created using the openai-python library to the OpenAI backend. The agent also captures the prompts and responses used by the OpenAI API models.
The following metrics are reported at the application, tier, and node levels:
- Calls per minute
- Errors per minute
- Tokens
- Input Tokens
- Completion Tokens
The following are the example metric paths:
- Per model metric path -
Application Infrastructure Performance|OpenAI_Tier|OpenAI|ada|Tokens
- All model cumulative metrics path -
Application Infrastructure Performance|OpenAI_Tier|OpenAI|Tokens - All Models
- Net application level metrics path -
Overall Application Summary|OpenAI - Tokens - ada
OpenAI Moderation API
From version 24.2.0, Python Agent supports OpenAI Moderation API. The moderation API will prompt the users if the queried statement is policy compliant or not. The Python Agent captures the following metrics:
- Calls to moderations API: The number of times moderation API is used every minute. (
OpenAI|moderations|Calls per minute)
- Flagged category metrics: The number of times specific types of content, like hate speech, are flagged. (
OpenAI|moderations|Flagged Calls by category|hate)
- Flagged queries count: The total amount of flagged content. (
OpenAI|moderations|Flagged queries)
- Total queries count: The overall number of queries processed. (
OpenAI|moderations|Total queries)
Prerequisites
Ensure that you meet the following prerequisites to monitor the OpenAI API exit calls with AppDynamics:
- Install Python Agent 23.8.0 or later. See Install the Python Agent.
- Import custom dashboards and templates.
- Set the
enable-openai
flag to true in the/path/to/appdynamics.cfg
file of the Python Agent. See[instrumentation]
. - To capture OpenAI request and response in snapshot exit call detail, set the
enable-genai-data-capture
flag to true in the/path/to/appdynamics.cfg
file of the Python Agent. See[instrumentation]
.
Create a Custom Dashboard
Example Dashboard
Create a Custom Cost Metric Expression
You can create a custom cost metric expression by multiplying the Input Tokens and Completion Tokens metrics by their respective costs.
For example, the cost per one million tokens for gpt-3.5-turbo-0125 is:
- Input Tokens = 0.5$
- Completion Tokens = 1.5$
gpt-3.5-turbo-0125
to calculate the cost as follows:({input Tokens}*0.5 + {Completion Tokens}*1.5)/1000000
Use this custom expression to create a pie chart for cost distribution across different models or an overall cost widget that includes the cost function for all models. Complete the following steps to create a cost per model widget to the existing OpenAI Dashboard:

This metric expression includes the cost multiplier per million tokens for both types.