Alerts
Monitor your LLM operations with configurable alerts. Get notified when error rates spike, latency degrades, or costs exceed thresholds.
Overview
Demeterics alerts help you:
- Detect issues early before they impact users
- Control costs with spending thresholds
- Monitor performance with latency alerts
- Track reliability with error rate monitoring
Alert Types
Error Rate
Monitor the percentage of failed LLM requests.
| Field | Description |
|---|---|
metric |
error_rate |
warning_value |
Percentage (0-100) to trigger warning |
critical_value |
Percentage (0-100) to trigger critical |
evaluation_window_minutes |
Time window for calculation (default: 5) |
Example configuration:
| Setting | Value |
|---|---|
| Warning threshold | 5% error rate |
| Critical threshold | 10% error rate |
| Evaluation window | 5 minutes |
P99 Latency
Monitor the 99th percentile response time in milliseconds.
| Field | Description |
|---|---|
metric |
p99_latency_ms |
warning_value |
Milliseconds to trigger warning |
critical_value |
Milliseconds to trigger critical |
evaluation_window_minutes |
Time window for calculation (default: 5) |
Example configuration:
| Setting | Value |
|---|---|
| Warning threshold | 5000ms (5 seconds) |
| Critical threshold | 10000ms (10 seconds) |
| Evaluation window | 5 minutes |
Cost Ratio
Monitor the cost relative to a baseline or budget.
| Field | Description |
|---|---|
metric |
cost_ratio |
warning_value |
Multiplier of baseline (e.g., 1.5 = 150%) |
critical_value |
Multiplier of baseline |
Example configuration:
| Setting | Value |
|---|---|
| Warning threshold | 1.5x (150% of baseline) |
| Critical threshold | 2.0x (200% of baseline) |
Note: Low credit balance alerts are configured separately in Settings → Credits. See Credits & Pricing for auto-reload and balance alert configuration.
Notification Channels
Alerts are sent to the email addresses configured for your account.
Configuration:
{
"channel": "email",
"recipients": ["ops@example.com", "oncall@example.com"]
}
Slack
Send alerts to a Slack channel via incoming webhook.
Setup:
- Create a Slack app at https://api.slack.com/apps
- Add an Incoming Webhook
- Copy the webhook URL
- Add to Demeterics: Settings → Integrations → Slack
Configuration:
{
"channel": "slack",
"webhook_url": "https://hooks.slack.com/services/T00/B00/xxx"
}
Webhook
Send alerts to any HTTP endpoint.
Configuration:
{
"channel": "webhook",
"url": "https://your-server.com/alerts",
"headers": {
"Authorization": "Bearer your-token"
}
}
Webhook Payload:
{
"alert_id": "alt_abc123",
"alert_name": "High Error Rate",
"severity": "critical",
"metric": "error_rate",
"current_value": 12.5,
"threshold": 10,
"triggered_at": "2025-01-29T12:00:00Z",
"message": "Error rate is 12.5%, exceeding critical threshold of 10%"
}
Managing Alerts
Alerts are managed through the Demeterics dashboard. A REST API for programmatic alert management is planned for a future release.
Create Alert
- Navigate to Alerts in the main menu (or Settings → Alerts)
- Click Create Alert
- Select metric type (
error_rate,p99_latency_ms, orcost_ratio) - Set warning and critical thresholds
- Configure the evaluation window
- Choose notification channels (email, Slack, webhook)
- Click Save
Edit Alert
- Navigate to Alerts
- Click the alert you want to modify
- Update thresholds or notification channels
- Click Save
Enable/Disable Alert
Toggle the Enabled switch on any alert to pause or resume monitoring without deleting the configuration.
Delete Alert
- Navigate to Alerts
- Click the alert you want to remove
- Click Delete and confirm
Alert History
View triggered alerts and their resolution status in the dashboard.
Viewing History
Navigate to Alerts → History to see:
- All triggered alerts
- Severity level (warning or critical)
- Trigger time
- Resolution status
- Current value vs threshold
Acknowledging Alerts
When an alert triggers:
- Click Acknowledge to indicate you're investigating
- The alert remains visible but marked as acknowledged
- Once the issue is resolved, click Resolve
Alert Lifecycle
- Active - Condition is currently met, notifications sent
- Acknowledged - Someone is investigating
- Resolved - Condition no longer met or manually resolved
Best Practices
Start Conservative
Begin with higher thresholds and tighten them as you understand your baseline:
- Set warning threshold at 2x your normal baseline
- Set critical threshold at 3x your normal baseline
- Monitor for 1-2 weeks
- Adjust based on actual alert frequency
Avoid Alert Fatigue
Too many alerts lead to ignored alerts:
- Use warning for investigation, critical for action
- Set appropriate time windows (5-15 minutes)
- Group related alerts
- Review and tune thresholds monthly
Use Multiple Channels
For critical alerts, configure redundant notification channels. In the alert configuration, enable multiple channels (email + Slack + webhook) to ensure notifications are received even if one channel fails.
Test Your Alerts
Verify alerts work before you need them:
- Create a test alert with a low threshold
- Trigger the condition intentionally
- Verify notification delivery
- Reset to production thresholds
Troubleshooting
Alerts Not Triggering
- Verify alert is enabled (Settings → Alerts)
- Check threshold values are appropriate
- Confirm time window covers recent activity
- Verify there's traffic to monitor
Not Receiving Notifications
- Check email spam folder
- Verify Slack webhook URL is correct
- Test webhook endpoint independently
- Check notification channel configuration
Too Many Alerts
- Increase thresholds
- Extend time window
- Use hysteresis (require sustained condition)
- Review baseline metrics
What's Next?
- Credits & Pricing - Set up low balance alerts
- API Reference - Full API documentation
- Dashboard - View real-time metrics