Why Observability Matters
LLM applications are inherently complex and non-deterministic. Without proper observability, you’re flying blind:- Debug production issues - Understand why specific requests failed or produced unexpected results
- Track agent workflows - Visualize multi-step AI agent processes from start to finish
- Monitor performance - Identify bottlenecks, latency issues, and cost spikes
- Analyze user behavior - See how users interact with your AI features
- Optimize costs - Track spending by user, feature, or workflow to control expenses
Core Observability Features
Requests
View and query every LLM request with full request/response bodies, metadata, and performance metrics
Sessions
Group related requests into sessions to trace complete AI agent workflows and multi-turn conversations
Traces
Log custom traces for non-LLM operations like database queries, API calls, and tool executions
Metrics
Analyze aggregate metrics across requests, sessions, and users to understand system-wide performance
Custom Properties
Tag requests with metadata for filtering, segmentation, and cost analysis by any dimension
User Metrics
Track per-user costs, usage patterns, and engagement metrics
How It Works
Helicone’s observability system operates at multiple levels:Request Level
Every LLM request flows through Helicone, capturing:- Complete request and response bodies
- Token counts and cost calculations
- Latency and time-to-first-token metrics
- Provider, model, and status information
- Custom properties and metadata
Session Level
Related requests are grouped using session headers:- Track multi-step agent workflows
- Visualize parent-child request relationships
- Analyze session-level metrics (total cost, duration, request count)
- Debug complex conversation flows
Custom Traces
Log non-LLM operations to get complete visibility:- Database queries and vector searches
- API calls and external tool executions
- Custom business logic and data processing
- Any operation you want to track within your workflow
Quick Start
View in Dashboard
Visit helicone.ai/requests to see your data:
- Filter by properties, model, date range, or status
- Click into individual requests to see full details
- View sessions to trace multi-step workflows
- Analyze metrics and trends over time
Common Observability Patterns
Debugging Production Issues
Tracking Agent Workflows
Cost Monitoring by Feature
Data Retention
Helicone stores your observability data based on your plan:- Free Plan: 30 days of request data
- Pro Plan: 90 days of request data
- Enterprise Plan: Custom retention (up to unlimited)
Querying Your Data
Access observability data programmatically via REST API:Query Requests
Filter and export request data for analysis
Query Sessions
Retrieve session data with all related requests
User Metrics
Analyze per-user usage and costs
Custom Exports
Export large datasets using our CLI tool
Advanced Features
Real-time Monitoring
- Live request feed in the dashboard
- Webhook notifications for specific events
- Alerts for errors, rate limits, or cost thresholds
Performance Analysis
- Latency percentiles (p50, p95, p99)
- Time-to-first-token tracking
- Request rate and throughput metrics
- Error rate monitoring by provider and model
Cost Optimization
- Cost breakdowns by model, user, and feature
- Token usage analysis and optimization suggestions
- Budget alerts and spending limits
- Cache hit rate tracking for cost savings
Privacy & Security
Helicone takes data privacy seriously:- Encryption: All data encrypted in transit (TLS) and at rest
- Isolation: Each organization’s data is isolated
- Access Control: Role-based access control for teams
- Compliance: SOC 2 Type II compliant, GDPR ready
- Data Residency: EU region available for GDPR compliance
Next Steps
View Requests
Explore the Requests page and learn how to query your data
Track Sessions
Group related requests into sessions for workflow tracking
Add Custom Properties
Tag requests with metadata for filtering and analysis
Analyze Metrics
Understand your system’s performance with aggregate metrics
Questions?
Need help or have questions? We’re here to help:- Discord Community: Join our Discord server for quick help
- GitHub Issues: Report bugs or request features on GitHub
- Documentation: Check our full documentation for more guides
