Skip to main content

Documentation Index

Fetch the complete documentation index at: https://prismeai-docs-next.mintlify.app/llms.txt

Use this file to discover all available pages before exploring further.

Observability Dashboard
Observability provides deep visibility into your AI workspaces, tracking performance metrics, error rates, LLM costs, and detailed request traces. Access it from Observability in the Governe sidebar.

Dashboard Overview

The observability dashboard shows key metrics at a glance:
MetricDescription
Health Score0-100 composite score based on errors, latency, and availability
Total ExecutionsNumber of automation runs in the selected period
Error RatePercentage of executions that failed
P95 Latency95th percentile response time
LLM CostsTotal cost of LLM calls

Health Score

The health score combines multiple factors into a single 0-100 metric:
StatusScore RangeColor
Healthy90-100Green
Degraded70-89Yellow
Critical0-69Red

Factor Breakdown

FactorWeightCalculation
Error Rate40%Lower error rate = higher score
Latency30%Lower latency = higher score
Availability30%Higher uptime = higher score

Performance Metrics

Latency Percentiles

Track response time distribution:
PercentileDescription
P50Median response time
P9090% of requests faster than this
P9595% of requests faster than this
P9999% of requests faster than this

Slowest Automations

Identify bottlenecks by viewing automations ranked by average execution time.

Timeline View

Visualize performance over time with:
  • P95 latency trend (area chart)
  • Request volume (line overlay)
This helps correlate load spikes with latency increases.

Error Tracking

Error Distribution

View errors grouped by:
  • Type: InvalidExpressionSyntax, FetchError, TimeoutError, etc.
  • Automation: Which automations generate the most errors

Error Rate Trend

Track error rate changes over time to identify:
  • New bugs introduced
  • External service outages
  • Configuration issues

Recent Errors

View the latest errors with:
  • Timestamp
  • Error type and message
  • Automation name
  • Correlation ID (for tracing)

LLM Costs

Monitor AI model consumption and costs:

Cost Breakdown

MetricDescription
Total CallsNumber of LLM API calls
Total CostEstimated cost in USD
Total TokensInput + output tokens consumed
Carbon ImpactEstimated CO2 emissions (kg)

By Model

See which models consume the most resources:
  • Model name
  • Number of calls
  • Cost percentage
  • Token count

By Automation

Identify which automations drive LLM costs:
  • Automation name
  • Call count
  • Relative usage

Cost Timeline

Track daily cost trends to:
  • Forecast monthly spend
  • Identify cost anomalies
  • Plan capacity

Request Traces

Trace individual requests through the entire execution flow.

Accessing Traces

  1. From the Observability dashboard, click View Traces
  2. Or click a correlation ID in the error list

Trace View

Each trace shows:
ElementDescription
SummaryEvent count, duration, user info
TimelineWaterfall view of all events
AutomationsList of automations executed
ErrorsAny errors encountered

Timeline Visualization

The waterfall chart shows:
  • Each automation as a horizontal bar
  • Start time offset from request start
  • Duration of each step
  • Errors highlighted in red

Dependency Graph

Visualize how automations call each other.

Graph View

  • Nodes: Automations and installed apps
  • Edges: Calls between components
  • Edge weight: Call frequency

Use Cases

  • Understand system architecture
  • Identify critical paths
  • Find orphaned automations

Anomaly Detection

Automatic detection of unusual patterns.

Anomaly Types

TypeDescription
Error SpikeSudden increase in error rate
Latency SpikeSudden increase in response time
Traffic SpikeUnusual increase in request volume
Traffic DropUnusual decrease in activity

Severity Levels

SeverityThreshold
Warning2x baseline
Critical3x baseline

Configuration

Adjust detection sensitivity:
  • Low: Only alert on significant anomalies
  • Medium: Balanced detection (default)
  • High: Alert on minor deviations

Period Comparison

Compare metrics between two time periods.

How to Compare

  1. Select current period dates
  2. Select comparison period dates
  3. View side-by-side metrics

Compared Metrics

  • Execution count
  • Error count and rate
  • Average and P95 latency
  • Unique users

Trend Indicators

  • Up: Increase > 5%
  • Down: Decrease > 5%
  • Stable: Within ±5%

Real-Time Feed

Monitor events as they happen.

Event Types

TypeDescription
runtime.automations.executedAutomation completed
runtime.fetch.completedHTTP fetch succeeded
runtime.fetch.failedHTTP fetch failed
runtime.llm.completedLLM call completed
errorGeneral error event

Filtering

  • All: Show all events
  • Errors: Only error events
  • Executions: Only completion events

Time Window

Configure how far back to show events (default: 60 minutes).

Time Range Selection

All views support time range selection:
PeriodDescription
1hLast hour
24hLast 24 hours
7dLast 7 days
30dLast 30 days
CustomSpecific date range

Platform Summary (Admin Only)

Platform administrators see a platform-wide summary:
  • Aggregate health score across all workspaces
  • Total platform requests
  • Combined LLM costs
  • Peak capacity usage

Best Practices

Set Baselines

Establish normal performance baselines for comparison

Monitor Anomalies

Enable anomaly detection for early warning

Track Costs

Review LLM costs weekly to prevent surprises

Use Traces

Investigate errors using correlation IDs

Troubleshooting

  1. Check Slowest Automations for bottlenecks
  2. Review the Timeline for load correlation
  3. Examine traces for slow steps
  4. Check external API response times

Next Steps

Model Governance

Control model access and costs

Audit Logs

Track administrative changes