From local development to enterprise deployment
SDK
Instrument & Capture
Cloud
Store & Analyze
Team
Collaborate & Debug
Auto-detect and instrument all major LLM providers and agent frameworks with a single prela.init() call
OpenAI
Chat, completions, embeddings, streaming
Anthropic
Messages API, streaming, tool use
LangChain
Chains, agents, tools, retrievers
LlamaIndex
Query engines, retrieval, synthesis
CrewAI
Task delegation, crew coordination
AutoGen
Conversational agents, function calling
LangGraph
Graph-based execution, state tracking
Swarm
Handoff-based agent switching
n8n webhook
HTTP-based workflow tracing
n8n code node
Python helpers for in-workflow tracing
Capture every LLM call, tool invocation, retrieval query, and agent handoff
Add observability without slowing down your agents
<1%
Latency Overhead
<100ms
Query Performance
10+
Frameworks Supported
Latency Impact Comparison
From structural checks to semantic similarity - assert on any aspect of agent behavior
containsText substring matchequalsExact matchregexPattern matchinglatencyExecution time checktoken_countToken usage validationtool_calledBasic tool use checktool_countNumber of tools calledno_piiPII detection assertionno_injectionInjection detection assertioncustom_ruleCustom regex/callable rulejson_schemaValidate against JSON schematool_sequenceVerify tool call orderagent_usedMulti-agent participationtask_completedTask completion verificationdelegation_occurredDelegation checkhandoff_occurredAgent handoff verificationsemantic_similarityEmbeddings-based comparisonnode_completedn8n workflow node completionworkflow_durationn8n execution timeai_node_tokensn8n AI node token validationllm_judgeAI-scored evaluation with rubricsCapture full execution context and replay with different models, parameters, or tool behavior
3 Report Formats
Console
Rich output for local development
JSON
Machine-readable for automation
JUnit
CI/CD integration
Run evaluation suites in your CI pipeline with JUnit reporter
GitHub Actions Example
WebSocket-based real-time updates with <1ms notification latency
Specialized dashboards for different agent architectures
This trace shows the hallucination issue. Check the grounding score.
Fixed! Replay shows 98% grounding now. Deploying.
Share traces, add comments, invite teammates
Natural language search and advanced filtering
"show me traces where gpt-4 timed out"model=gpt-4, status=error, duration>5slast 1h, 24h, 7d, 30dtag:production, tag:stagingspend>$1Export Capabilities
CSV Export
Lunch Money+
JSON Export
Lunch Money+
Programmatic API
Pro+ • Full API access
Automatically detect when agents make claims not grounded in source documents
⚠️ Detected Hallucination
"The Eiffel Tower was built in 1920"
Source documents state: built in 1889
Baseline tracking with automated anomaly alerts
Response length +42% in last 2 days
AI-powered recommendations to reduce costs without sacrificing quality
💰 Monthly Savings Report
$847
Saved this month
23%
Cost reduction
Categorize errors and provide one-click fix suggestions
Rate limit exceeded
→ Try gpt-4o-mini (83% cheaper, higher quota)
Token limit
→ Increase max_tokens by 50% (+$0.15/call)
Auth failure
→ Check API key in settings
Network error
→ Retry with exponential backoff
Real-time input/output filtering with configurable actions — block, redact, or log
AI-scored evaluations using custom rubrics with threshold-based pass/fail
Example Rubric
criteria: "Is the response helpful and accurate?"
threshold: 0.7
model: claude-sonnet-4-20250514
Version-controlled prompt templates with stage-based promotion
Classify: {{text}} → Categories: {{categories}}
Event-driven microservices architecture
<100ms
Simple query performance
<500ms
Complex aggregations
90-day
TTL with monthly partitioning
Horizontal
Scalability