You just deployed a machine learning model on Google’s Vertex AI. It’s scaling beautifully until your ops dashboard lights up like a warning beacon. Latency spikes, memory climbs, and a few thousand logs later you still can’t spot the root cause. That’s where the New Relic Vertex AI integration changes the story from detective work to real-time insight.
New Relic gives you observability: distributed tracing, metrics, and application performance data. Vertex AI brings model training, prediction, and managed pipelines inside Google Cloud. Together, they let teams stitch AI behavior into the same operational lens you already use for code and infrastructure. No more hunting across dashboards to guess whether the slowdown came from your model or the network edge.
When you link New Relic to Vertex AI, telemetry moves from the AI platform straight into your observability stack. Each deployed model, endpoint, and prediction request appears like any other service. You see latency distributions, request counts, and inference errors, mapped against CPU and GPU utilization. That correlation is what turns raw data into a story you can act on.
How do you connect New Relic and Vertex AI?
Configuration happens at the project level. Vertex AI exposes metrics through Google Cloud Monitoring, which you can forward to New Relic via an integration key and the GCP exporter. Once authorized, New Relic treats your Vertex resources like first-class citizens. Identity and access control still flow through IAM or OIDC, so your SOC 2 and RBAC boundaries remain intact. In short, use the same service accounts and policies you trust elsewhere.
For best results, use a naming convention that matches model endpoints to production services. Rotate integration keys like any secret. Audit logs regularly so you know which automation or service account is writing telemetry data. You don’t need to see every request, only the ones that reveal patterns over time.