Skip to main content1. Description
The Dashboard module in Bud AI Foundry gives teams a single-pane overview of platform activity across cloud and local models (Hugging Face, URL-based, and disk-based), deployments, projects, and clusters. It highlights request volumes, throughput, latency, token usage, and accuracy for the top models.
Each chart carries its own time-range toggle (Last 24 Hrs, Last 7 Days, Last 30 Days) so platform, MLOps, and FinOps teams can compare day-over-day, week-over-week, or month-over-month changes without leaving the overview.
2. USPs (Unique Selling Propositions)
1. Hybrid-aware overview for cloud and local models
Shows cloud and local (Hugging Face, URL, disk) model counts together with active endpoint and cluster readiness, reflecting Bud’s CPU-first, burst-to-cloud strategy.
2. Time-aligned deltas across every metric
Segmented controls on each chart expose Last 24 Hrs, Last 7 Days, or Last 30 Days windows, pairing totals with delta percentages for quick trend checks.
3. Top-model intelligence with accuracy and tokens
Highlights the top five models by usage, accuracy benchmarks (MMLU, ARC-C, GSM8K, Math, HumanEval), and input/output token mix so you can spot quality and cost shifts early.
4. Operational posture cards
Instant readiness signals for endpoints, clusters, and projects (running vs inactive counts, member totals) so you can triage where to intervene before scaling or routing changes.
5. Navigation that mirrors downstream workflows
Sidebar tabs for Projects, Models, Clusters, Playground, API Keys, Agents, Observability, and more let you jump from the overview into detail pages and take actions without losing context.
3. Features
3.1 Landing layout and navigation
- Dashboard opens inside the unified Bud Admin layout with sidebar tabs for Projects, Models, Clusters, Dashboard, Playground, API Keys, Agents, Observability, and optional dev-only entries.
- Carousel hero card for business highlights plus quick access to status cards and charts.
3.2 Total Requests card
- Displays last-7-day total requests with formatted values (K/M/B) and delta percentage (green up arrow for positive, red down for negative).
- Inline sparkline shows request trajectory for the past week (daily points).
3.3 Model totals and breakdown
- Shows total models with separate chips for Cloud and Local counts, covering cloud endpoints and local models (Hugging Face, URL, disk-based).
3.4 Endpoint, Cluster, and Project readiness cards
- Endpoints: total plus running counts.
- Clusters: total plus inactive counts.
- Projects: total plus member counts.
- Cards emphasize readiness colors (green for healthy, red for issues, blue for collaboration).
3.5 API Calls chart
- Bar chart of API calls grouped by project with delta badge showing average change for the selected range.
- Time filters: Last 24 Hrs, Last 7 Days, Last 30 Days via Segmented control.
- Empty states show guidance when data is unavailable.
3.6 Latency chart
- Bar chart for latency by project with avg delta chip; values shown in milliseconds.
- Time filters: Last 24 Hrs, Last 7 Days, Last 30 Days.
3.7 Throughput chart
- Bar chart for throughput (tokens/s) by project with avg delta chip.
- Time filters: Last 24 Hrs, Last 7 Days, Last 30 Days.
3.8 Model Usage chart
- Bar chart for top five models by usage with avg delta chip.
- Time filters: Last 24 Hrs, Last 7 Days, Last 30 Days.
3.9 Accuracy leaderboard
- Accuracy chart for the top five models across benchmarks (MMLU, ARC-C, GSM8K, Math, HumanEval).
- Shows comparative bars for each benchmark per model; “For the top 5 models” caption clarifies scope.
3.10 Token Metrics
- Comparative bar chart for input vs output tokens across the top five models.
- Time filters: Last 24 Hrs, Last 7 Days, Last 30 Days.
- Delta chips highlight average changes for input and output tokens.
4. How-to Guides
4.1 Access the Dashboard
- Log in to Bud AI Foundry with SSO or credentials.
- Click Dashboard from the left navigation.
- Confirm the landing view shows the carousel, total requests, models breakdown, and status cards.
4.2 Adjust the time range per chart
- In a chart header, use the Segmented control to choose LAST 24 HRS, LAST 7 DAYS, or LAST 30 DAYS.
- The dashboard refreshes the selected chart using the matching frequency and delta window; repeat for other charts as needed.
4.3 Review total requests trend
- Locate the Total Requests card in the upper grid.
- Read the formatted count and delta percentage for the last 7 days (week-over-week delta).
- Use the sparkline (daily points) to see the past week’s trajectory; open Observability for per-request detail.
4.4 Inspect API calls by project
- Open the API Calls chart.
- Pick a time range via the Segmented control.
- Review bar heights per project and the Avg. delta chip to spot increases or regressions.
4.5 Monitor latency and throughput
- Open the Latency and Throughput charts.
- Select a time window to update values.
- Use the Avg. delta chips (green/red) to decide whether to route traffic, scale clusters, or inspect Observability for slow projects.
4.6 Track model usage and token mix
- Open Model Usage to see the top five models by usage for the selected window.
- Switch to Token Metrics to compare input vs output tokens for the same top models.
- Use the delta chips to judge cost/efficiency trends, especially for local deployments vs cloud endpoints.
4.7 Compare model accuracy
- Open Accuracy and confirm “For the top 5 models” caption.
- Review benchmark bars (MMLU, ARC-C, GSM8K, Math, HumanEval) per model.
- Use insights to adjust routing weights, evaluation plans, or guardrails for cloud and local variants.
4.8 Interpret readiness cards
- Scan Endpoints, Clusters, and Projects cards for totals and other details.
- Use running vs inactive counts to prioritize cluster remediation or endpoint restarts before traffic surges.
5. FAQ
5.1 Which models are counted in the dashboard?
Cloud endpoints and local models (Hugging Face downloads, URL-based imports, and disk-mounted checkpoints) are counted together.
5.2 What time windows are supported?
Each chart supports Last 24 Hrs (daily), Last 7 Days (weekly), and Last 30 Days (monthly) presets; deltas compare against the previous equivalent window.
5.3 How are the top five models chosen?
Charts request top_k=5 from analytics APIs, grouping by model for usage, token metrics, and accuracy benchmarks.
5.4 How do I navigate from the dashboard to take action?
Use the sidebar tabs to open Projects, Models, Clusters, Observability, or Agents. From there, drill into deployments, request traces, or routing policies and return to the dashboard for updated rollups.
5.5 Where can I see deeper analytics beyond the dashboard?
Use the Observability module for request-level traces, latency and token breakdowns, benchmarking comparisons across hardware and backends, and regressions after model upgrades. The dashboard surfaces rollups while Observability provides the detailed analytics and diagnostics. Refer to the [Observability guide] for more.