AI Performance

Learn how to use Sentry's AI Performance tool to get insights into things that may be affecting your application health, including critical LLM metrics.

Sentry's AI Performance allows you to see critical metrics for your LLM pipelines, like token usage and duration. In addition to having a dedicated space to monitor AI performance, you can also look view Sentry's Frontend, Backend, and Mobile Insights to understand the performance of the applications running your pipelines.

Was this helpful?
Help improve this content
Our documentation is open source and available on GitHub. Your contributions are welcome, whether fixing a typo (drat!) or suggesting an update ("yeah, this would be better").