Skip to content

Latest commit

 

History

History
310 lines (251 loc) · 10.5 KB

File metadata and controls

310 lines (251 loc) · 10.5 KB

TestOps Copilot - Demo & Screenshots

Visual guide to TestOps Copilot's features and user interface. Updated for v3.0.0 — Graduated Autonomy, Global AI Context, High-Fidelity Seeding.


Quick Start Demo

npm install && npm run dev:simple
# Login with any demo account (password: demo123):
# admin@testops.ai | lead@testops.ai | engineer@testops.ai | viewer@testops.ai

Everything below works in demo mode with zero external dependencies.


3-Column Mission Control Layout

+------------------+-----------------------------+---------------------+
|                  |                             |                     |
|  Navigation      |  Main Content               |  AI Copilot Panel   |
|  Sidebar         |  (Dashboard, KB,            |  (Chat + Persona    |
|                  |   Pipeline, Teams)           |   Badge + Cards)   |
|  - Dashboard     |                             |                     |
|  - Pipelines     |  +--- Stats Cards ---+      |  [Test Engineer is  |
|  - Test Runs     |  | Runs | Pass% | Flaky |  |   handling this]    |
|  - Knowledge Base|  +-------------------+      |                     |
|  - Teams         |                             |  > Analyzing your   |
|  - Settings      |  +--- Trend Chart ---+      |    question...      |
|  - Cost Tracker  |  |                   |      |                     |
|                  |  +-------------------+      |  [tool: dashboard]  |
|                  |                             |  Found 3 flaky tests|
|                  |  +--- Recent Runs ---+      |                     |
|                  |  | Run #423 FAILED   |      |  Based on the data, |
|                  |  | Run #422 PASSED   |      |  here are the top   |
|                  |  +-------------------+      |  flaky tests...     |
+------------------+-----------------------------+---------------------+

AI Copilot Workflows

Workflow 1: Investigating a Flaky Test

sequenceDiagram
    participant User
    participant Copilot as AI Copilot
    participant Router as PersonaRouter
    participant Tool as dashboard_metrics

    User->>Copilot: "Why are my login tests flaky?"
    Copilot->>Router: Classify query
    Router-->>Copilot: Test Engineer (keyword: "flaky")
    Note over Copilot: persona_selected SSE event
    Copilot->>Tool: dashboard_metrics({type: "overview"})
    Tool-->>Copilot: 3 flaky tests found
    Copilot-->>User: "As the Test Engineer, I found 3 flaky tests..."
Loading

What the user sees:

  1. Persona badge: "Test Engineer is handling this"
  2. Thinking indicator: "Analyzing your question..."
  3. Tool call: "Calling dashboard_metrics..."
  4. Rich result card with test data
  5. Markdown answer with analysis and recommendations

Workflow 2: Creating a Jira Issue (Human-in-the-Loop)

sequenceDiagram
    participant User
    participant Copilot as AI Copilot
    participant Confirm as Confirmation System
    participant Jira as jira_create_issue

    User->>Copilot: "Create a bug ticket for the login timeout"
    Copilot->>Copilot: PersonaRouter -> Senior Engineer
    Copilot-->>User: confirmation_request SSE event
    Note over User: Preview card with issue details
    Note over User: [Approve] [Deny] buttons (5-min TTL)
    User->>Confirm: Approve
    Confirm->>Jira: Execute tool
    Jira-->>User: Jira issue PROJ-789 created
    Copilot-->>User: "Issue created. Here's a summary..."
Loading

What the user sees:

  1. Preview card showing the Jira issue that will be created
  2. Approve/Deny buttons with a 5-minute countdown
  3. After approval: result card with the created issue details
  4. AI summarizes what was done and suggests next steps

Workflow 3: Cross-Platform Context Enrichment

sequenceDiagram
    participant User
    participant Copilot as AI Copilot
    participant Jira
    participant Confluence
    participant GitHub

    User->>Copilot: "What caused the connection timeout in test run 423?"
    Copilot->>Copilot: PersonaRouter -> Test Engineer
    par Parallel tool calls
        Copilot->>Jira: jira_search("connection timeout")
        Copilot->>GitHub: github_get_commit(sha)
        Copilot->>Confluence: confluence_search("timeout runbook")
    end
    Jira-->>Copilot: PROJ-456 (In Progress)
    GitHub-->>Copilot: PR #123 modified timeout config
    Confluence-->>Copilot: "Connection Troubleshooting" runbook
    Copilot-->>User: Synthesis with all context linked
Loading

Virtual Team Persona Routing

Every query is classified before the AI responds. The persona badge appears in the chat:

+-----------------------------------------------+
|  AI Copilot                      [Provider] [x] |
+-----------------------------------------------+
|                                                 |
|  You: "Why are my tests flaky?"                 |
|                                                 |
|  +----- [Test Engineer is handling this] -----+ |
|  |  Analyzing your question...                | |
|  |                                            | |
|  |  [Action: dashboard_metrics]               | |
|  |  Found 3 flaky tests in last 7 days       | |
|  |                                            | |
|  |  Based on the dashboard data, here are     | |
|  |  the top flaky tests:                      | |
|  |  1. test_login_timeout (68% pass rate)     | |
|  |  2. test_payment_flow (72% pass rate)      | |
|  |  3. test_api_health (85% pass rate)        | |
|  |                                            | |
|  |  **Recommendations:**                      | |
|  |  - Add retry logic for network calls       | |
|  |  - Increase timeout in test config         | |
|  +--------------------------------------------+ |
|                                                 |
|  [Type a message...]                    [Send]  |
+-------------------------------------------------+

Routing examples:

You Ask Persona Selected Badge Shows
"Why are my tests flaky?" TEST_ENGINEER "Test Engineer is handling this"
"Pipeline build failed" DEVOPS_ENGINEER "DevOps Engineer is handling this"
"What can this tool do?" AI_PRODUCT_MANAGER "Product Manager is handling this"
"Check for auth vulnerabilities" SECURITY_ENGINEER "Security Engineer is handling this"
"Database migration failed" DATA_ENGINEER "Data Engineer is handling this"
"API is slow" PERFORMANCE_ENGINEER "Performance Engineer is handling this"
"Search Jira for AUTH-123" SENIOR_ENGINEER "Senior Engineer is handling this"

Confirmation Cards (Human-in-the-Loop)

Write operations show a preview card before executing:

+-----------------------------------------------+
|  Jira Create Issue                             |
|                                                |
|  Project: PROJ                                 |
|  Type: Bug                                     |
|  Summary: Login timeout in E2E test suite      |
|  Priority: High                                |
|  Assignee: Auto                                |
|                                                |
|  [Approve]  [Deny]        Expires in 4:32      |
+-----------------------------------------------+

Write tools that require confirmation:

  • jira_create_issue -- Create Jira ticket
  • jira_transition_issue -- Change issue status
  • jira_comment -- Add comment to issue
  • github_create_pr -- Open pull request
  • github_create_branch -- Create branch
  • github_update_file -- Modify file contents

Dashboard Overview

graph TD
    subgraph Dashboard["Main Dashboard"]
        Stats[Statistics Cards]
        Charts[Test Trends and Charts]
        Recent[Recent Test Runs]
    end

    Stats --> Card1[Total Tests: 1,245]
    Stats --> Card2[Pass Rate: 94.2%]
    Stats --> Card3[Flaky Tests: 12]
    Stats --> Card4[Active Failures: 8]

    Charts --> LineChart[7-Day Pass/Fail Trend]
    Charts --> PieChart[Failure Categories]

    Recent --> Run1[Test Run #423 - FAILED]
    Recent --> Run2[Test Run #422 - PASSED]
Loading

Failure Knowledge Base

graph TD
    subgraph KB["Failure Knowledge Base"]
        Search[Search and Filters]
        Stats2[Quick Statistics]
        FailureList[Archived Failures]
        Predictions[Predictive Analysis]
    end

    Stats2 --> Total[Total: 1,600+]
    Stats2 --> Documented[RCAs: 298]
    Stats2 --> Recurring[Recurring: 23]

    Predictions --> Trends[Failure Trends Chart]
    Predictions --> Risk[Risk Score Table]
    Predictions --> Anomaly[Anomaly Detection]

    FailureList --> F1[DB Connection Timeout]
    F1 --> RCA[RCA Available + 3 Comments]
    F1 --> Revisions[2 Revisions + Activity Feed]
Loading

Key Feature Screenshots Checklist

Essential:

  • 3-column Mission Control layout (full view)
  • AI Copilot chat with persona badge
  • Persona routing in action (different queries, different badges)
  • Confirmation card with Approve/Deny buttons
  • Tool result card (e.g., dashboard_metrics)
  • In-chat provider picker dropdown
  • Failure Knowledge Base with trends chart
  • Team Workspaces management page
  • RCA modal with version history and comments

Copilot Interactions:

  • Flaky test investigation flow
  • Jira issue creation with confirmation
  • Cross-platform context enrichment results
  • Pipeline status check
  • "What can this do?" product discovery flow

Video Demo Suggestions

1. Quick Start Demo (2 minutes)

  • npm run dev:simple then browser opens
  • Login then see dashboard with populated data
  • Open AI copilot then ask about flaky tests
  • See persona badge + tool calls + analysis

2. AI Copilot Deep Dive (3 minutes)

  • Ask different queries, show different persona selections
  • Demonstrate a write operation with confirmation flow
  • Show cross-platform enrichment (Jira + Confluence + GitHub)
  • Switch AI provider mid-session with the picker

3. Failure Investigation (3 minutes)

  • Click into a failed test run
  • View the RCA document with revision history
  • Add a comment, see the activity feed
  • Use predictive analysis (trends, risk scores)

4. Team Workspaces (2 minutes)

  • Create a team, invite members
  • Assign pipelines to the team
  • Configure a shared dashboard

Branding

Element Value
Primary Blue #1976d2
Success Green #2e7d32
Error Red #d32f2f
Warning Orange #ed6c02
Font Roboto (headers bold, body regular)
Icons Material-UI throughout

This file contains visual mockups. Actual screenshots will be added as the application evolves. Contributors welcome!