Configure your Splunk Observability Cloud account to collect GCP VertexAI metrics
Learn how to configure your Splunk Observability Cloud account to collect GCP VertexAI metrics.
You can monitor the performance of Google Cloud Platform (GCP) VertexAI applications by configuring your GCP VertexAI applications to send metrics to Splunk Observability Cloud. This solution creates a cloud connection in your Splunk Observability Cloud account that collects metrics from Google Cloud Monitoring.
Complete the following steps to collect metrics from GCP VertexAI.
- Connect GCP to Splunk Observability Cloud. For more information on the connection methods and instructions for each method, see Connect to Google Cloud Platform.
- To monitor GCP VertexAI metrics with Splunk Observability Cloud, run your applications that use GCP VertexAI models.
Metrics
Learn about the available metrics for GCP VertexAI.
| Metric name | Unit | Description |
|---|---|---|
prediction/online /prediction_count | count | Number of online predictions. |
prediction/online /prediction_latencies | ms | Online prediction latency of the deployed model. |
prediction/online /response_count | count | Number of different online prediction response codes. |
prediction/online /prediction_latencies.count | count | Number of online predictions. |
prediction/online /prediction_latencies.sumOfSquareDeviation | ms | The sum of squared deviation for prediction latencies. |
publisher/online_serving /model_invocation_count | count | Number of model invocations (prediction requests). |
publisher/online_serving /model_invocation_latencies.sumOfSquareDeviation | ms | The sum of squared deviation for model invocation latencies. |
publisher/online_serving /model_invocation_latencies.count | count | Number of model invocations (prediction requests). |
publisher/online_serving /model_invocation_latencies | ms | Model invocation latencies (prediction latencies). |
publisher/online_serving /token_count | count | Accumulated input/output token count. |
publisher/online_serving /consumed_token_throughput | count | Overall throughput used (accounting for burndown rate) in terms of tokens. |
publisher/online_serving /consumed_throughput | count | Overall throughput used (accounting for burndown rate) in terms of characters. |
publisher/online_serving /character_count | count | Accumulated input/output character count. |
publisher/online_serving /first_token_latencies | ms | Duration from request received to first token sent back to the client. |
publisher/online_serving /first_token_latencies.count | count | Number of first token latencies. |
publisher/online_serving /first_token_latencies.sumOfSquareDeviation | ms | The sum of squared deviation for first token latencies. |
Attributes
Learn about the available resource attributes for GCP VertexAI.
gcp_project_statusgcp_project_namegcp_project_label_last_revalidated_bymodel_user_idgcp_project_numberrequest_typegcp_idgcp_project_label_cloud_registration_idgcp_project_creation_timegcp_project_label_last_revalidated_atinput_token_sizeoutput_token_sizeproject_idmetricTypeDomaingcp_project_label_environmentpublishermonitored_resourcegcp_project_label_account_typegcp_project_label_owner_groupserviceLocation
In addition, the type resource attribute is available for the publisher/online_serving /token_count and publisher/online_serving /character_count metrics.
Troubleshoot
Learn how to get help if you can't see your data in Splunk Observability Cloud.
If you can't see your data in Splunk Observability Cloud, you can get help in the following ways:
Splunk Observability Cloud customers can submit a case in the Splunk Support Portal or contact Splunk Support.
Prospective customers and free trial users can ask a question and get answers through community support in the Splunk Community.
Next steps
How to monitor your AI components after you set up Observability for AI.
After you set up data collection from supported AI components to Splunk Observability Cloud, the data populates built-in experiences that you can use to monitor and troubleshoot your AI components.
| Monitoring tool | Use this tool to | Link to documentation |
|---|---|---|
| Built-in navigators | Orient and explore different layers of your AI tech stack. | |
| Built-in dashboards | Assess service, endpoint, and system health at a glance. | |
| Splunk Application Performance Monitoring (APM) service map and trace view | View all of your LLM service dependency graphs and user interactions in the service map or trace view. |