• Error Monitoring
  • Logs
  • Session Replay
  • Tracing
  • Seer
  • Uptime Monitoring
  • Profiling
  • Cron Monitoring
  • AI Code Review
  • Github
  • Slack
  • All Integrations
  • Javascript
  • Python
  • React
  • Laravel
  • Next.js
  • All SDKs

Solutions

  • Web / Frontend Development
  • Mobile Crash Reporting
  • Game Crash Reporting
  • AI Observability
  • Application Performance Monitoring
  • Real User Monitoring
  • Ecommerce
  • Enterprise
  • Blog
  • Changelog
  • Sandbox
  • Resources
  • Sentry Answers
  • Syntax
  • Customers
  • For Startups
  • Contact Us
  • Help Center
  • Status
  • Sentry Build
  • Events
  • Merch
Docs
Pricing
Sign InGet DemoGet Started
  • Products

    • Error Monitoring
    • LogsNEW
    • Session Replay
    • Tracing
    • SeerNEW
    • Uptime Monitoring
    • Profiling
    • Cron Monitoring
    • AI Code ReviewNEW

    Products

    • Error Monitoring
    • LogsNEW
    • Session Replay
    • Tracing
    • SeerNEW
    • Uptime Monitoring
    • Profiling
    • Cron Monitoring
    • AI Code ReviewNEW

    Integrations

    • Github
    • Slack
    • All Integrations

    Integrations

    • Github
    • Slack
    • All Integrations

    SDKs

    • Javascript
    • Python
    • React
    • Laravel
    • Next.js
    • All SDKs

    SDKs

    • Javascript
    • Python
    • React
    • Laravel
    • Next.js
    • All SDKs
  • Solutions

    • Web / Frontend Development
    • Mobile Crash Reporting
    • Game Crash Reporting
    • AI Observability
    • Application Performance Monitoring
    • Real User Monitoring
    • Ecommerce
    • Enterprise
  • Learn

    • Blog
    • Changelog
    • Sandbox
    • Resources
    • Sentry Answers
    • Syntax
    • Customers
    • For Startups

    Learn

    • Blog
    • Changelog
    • Sandbox
    • Resources
    • Sentry Answers
    • Syntax
    • Customers
    • For Startups

    Support

    • Contact Us
    • Help Center
    • Status

    Support

    • Contact Us
    • Help Center
    • Status

    Hang out with us

    • Sentry Build
    • Events
    • Merch

    Hang out with us

    • Sentry Build
    • Events
    • Merch
    Holiday E-Commerce Checklist: A Developer’s Survival Guide

    Holiday E-Commerce Checklist: A Developer’s Survival Guide

    There’s never a good time for errors or performance degradations to show up, but during periods of peak traffic like the holidays, it’s especially critical to get immediate answers about what's failing and how to fix it.

    Learn More
  • Docs
  • Pricing
Sign InGet DemoGet Started
Sentry Changelog Illustration

Sentry Changelog

Follow Twitter@SentryChangelog to stay up to date on everything from product updates to SDK changes.

Categories:

Jump to:

LLM Monitoring  (Beta)

LLM Monitoring (Beta)

Sentry LLM Monitoring offers organizations rich debugging context, and visibility into the cost and performance of applications calling LLMs. Developers get better visibility into their LLM-powered applications with automatic token cost and usage calculations for their LLMs. Developers can identify the root cause of an issue fast with detailed debugging context, like the model version, user prompts, and the sequence of calls to the LLM. With real-time visibility into errors, performance bottlenecks and cost, organizations can make their LLM-powered applications more effective and efficient.

llm-monitoring-ss.png

Sentry LLM Monitoring is for you if:

  • You want to understand the costs of your LLMs in production, and set alerts on cost overruns.
  • It's currently difficult to debug issues with your LLMs, especially subjective failures like bad responses.
  • You want to understand the performance of LLMs in the context of other operations like database calls and HTTP servers.

Core Features

  • Tracking token usage and cost across multiple LLM models in real-time.
  • Alerting on token usage and cost.
  • Detailed debugging context like the model version, prompt and the sequence calls made to the LLM.
  • Visibility of token usage, cost and performance by AI pipeline.

Getting started

Sentry LLM Monitoring is automatically enabled for users of the Sentry Python SDK. You can also use our provided decorators to manually create AI pipelines and store token usage information. llm-code.png

You'll need:

  • LangChain, or a Python function instrumented with @ai_track that calls one or more LLMs.
  • One of the official Python packages for OpenAI, Anthropic, Huggingface Hub, or Cohere, or explicit token usage measurements with the record_token_usage function.
June 11, 2024
Your code is broken. Let's Fix it.
Get Started