Skip to main content

AI Baseline (Adoption & Impact)

Overview

AI Baseline is your org's AI impact dashboard. It shows:

  • How widely AI coding tools (e.g., GitHub Copilot) are used and how adoption changes over time.
  • Side-by-side delivery outcomes “with AI” vs “without AI.”
  • Usage patterns by model, editor, and features.
  • License utilization to optimize spend.

Use it to validate AI ROI, spot adoption gaps, and prioritize enablement or license changes.

image

Who can access

Cloud • Available to users with Engineering Manager (and above) permissions.

What you'll learn

  • Adoption momentum: weekly trends of active AI users and acceptance of AI suggestions.
  • Delivery acceleration: changes in cycle time, review time, rework, and lead time with AI vs without AI.
  • Quality and predictability: change failure rate, bugs %, planned vs spillover, sprint completion.
  • Usage insights: most-used editors/models and feature usage; seats and activity to right-size licenses.

Prerequisites

  • Connected code provider:
    • GitHub (Gitlab, Cursor, etc. coming soon.)
  • Required scopes:
    • GitHub: read:org (for organization and license data).
  • Optional: AI user tagging
    • Tag who counts as an “AI user” to enable focused comparisons (AI-only vs Non-AI users).
Where to connect

Settings → Integrations → Code Provider. If prompted on the baseline UI, re-link to grant missing scopes.

Data sources

  • Code provider activity (GitHub/etc.) for PRs, reviews, and repository context.
  • AI usage events (e.g., code suggestions and acceptances) from the provider's AI interfaces where available.
  • Organization license information (provider-specific).
  • Active users and org settings (for AI user tagging).

Product → AI → Baseline

The page includes:

  • Header: date range, org selector (code provider org/group), user filter, and “Generate AI Analysis.”
  • Tabs:
    • AI Adoption & Trends
    • Usage Insights

Filters and controls

  • Date range: Focus the analysis window (weekly cadence).
  • Code provider group: Choose the GitHub organization or GitLab group to analyze.
  • Users filter:
    • None: All users.
    • AI users only (include): Only users tagged as AI users.
    • Non-AI users (exclude): Everyone except tagged AI users.
AI users tagging

Configure AI users in Settings → AI Users (or via the in-page “Configure AI Users” prompt). This enables apples-to-apples comparisons for velocity trends and speed metrics.

image

What's included:

  • Adoption momentum: Weekly view of active AI usage. When total users are known, the chart can show percentage of users engaging with AI.
  • AI adoption rate: The peak weekly AI usage count during the period, divided by total active users (capped at 100%).
  • Adoption rate change: Relative change from the first to the last week in the selected window.
  • AI usage trends: Weekly counts of AI suggestions and acceptances.
  • Delivery acceleration (With AI vs Without AI):
    • Cycle time
    • First response time (time to first review)
    • Rework time (iterations required)
    • Lead time, Merge time, Merge-to-deploy
    • PR count
  • Executive summary:
    • Speed: cycle time, first response time, rework time, PRs per developer.
    • Quality: change failure rate, bugs percentage, PRs merged without review.
    • Predictability: planned %, spillover %, sprint completion.

How “With AI” vs “Without AI” works

  • Default mode: Aggregated impact based on the provider's AI signals for your org.
  • AI-only/Non-AI-only modes: When an AI user filter is set, metrics are derived from the tagged user cohort (by mapping your tagged MHQ users to their code-host usernames).

Usage Insights

image

  • AI usage breakdown:
    • Model preference (e.g., Copilot, Gemini, Claude).
    • Editor distribution (e.g., VS Code, WebStorm).
    • Feature usage (e.g., chat, completions, code suggestions).
  • License utilization:
    • Total seats by org/group and activity recency.
    • Per-user last activity to identify waste or opportunities to reassign seats.

Metric definitions

More detailed definitions here: Link

  • AI adoption rate
    • Definition: Max weekly AI usage count during the selected period ÷ total active users (in MHQ), expressed as a percentage (0-100).
  • AI adoption rate change
    • Definition: Relative change between the first and last period's adoption rate, expressed as a percentage.
  • AI suggestions count / accepted
    • Weekly counts of suggestions shown vs accepted by developers.
  • Delivery metrics
    • Cycle time: Time from first commit to merge (days).
    • First response time: Time to first reviewer response (days).
    • Rework time: Time spent in rework between review cycles (days).
    • Lead time: Time from work start to merge (days).
    • Merge time: Time from PR open to merge (days).
    • Merge-to-deploy: Time from merge to production deployment (days).
    • PR count: Number of PRs in the window.
  • Executive summary
    • PRs merged without review: Count of PRs merged without at least one review.
    • Change failure rate: % of releases requiring remediation/rollback (org-specific).
    • Bugs percentage: Share of tickets flagged as bugs.
    • Planned percentage: % of tickets planned for the sprint/release.
    • Spillover percentage: % of tickets that slipped to a subsequent iteration.
    • Sprint completion: % completed in the planned window.
Units

Time metrics are shown in days (rounded for readability). Trends are aggregated weekly.

AI-generated summary

Click “Generate AI Analysis” to produce an executive-friendly summary of your org's adoption, trends, and impact, with actionable recommendations.

  • Availability: Feature-flagged. Requires permission to share chart data with the AI summarization service.
  • Output: Streamed narrative you can copy for status updates or reviews.

image

First-time walkthrough

The first time you land on Baseline, a short walkthrough highlights key insights and controls. Dismiss or revisit any time; your “seen” status is remembered per user.

Troubleshooting

  • “GitHub Integration Required”
    • Link your GitHub org under Settings → Integrations.
  • “Missing required scopes”
    • Re-link GitHub and grant the requested scopes (e.g., read:org).
  • “No orgs connected”
    • Ensure the integration has access to at least one organization/group.
  • “No AI activity detected yet”
    • You'll see insights once developers start using AI coding tools and the provider emits usage signals.
  • “No AI Review data”
    • If you're comparing “With AI Review” vs “Without,” ensure your teams use AI reviewers (e.g., GitHub Copilot PR review).

Permissions and roles

  • Minimum role: Engineering Manager (or higher).
  • Org admins can manage integrations, scopes, and AI user tagging.

Limitations

  • Data availability varies by provider and scope grants.
  • AI user comparisons depend on accurate AI user tagging and identity mapping to code-host usernames.
  • Some quality/predictability metrics depend on your project/ticketing integrations and configuration.

Best practices

  • Start with a recent 4 week window to establish a baseline.
  • Tag AI users to run clean comparisons, then expand tagging as adoption grows.
  • Use the executive summary for leadership updates; drill into trends to prioritize enablement.

Last updated: 2025-09-01