Graphite computes stats for each user as an author and as a reviewer. Stats are computed on merged PRs only.

Methodology

Active users: For a user to be active in the selected time frame and repos they must have:

  1. created a PR on GitHub to be considered active on GitHub or

  2. created a stack of 2+ PRs on Graphite to be considered active on Graphite

Rollups: The graphed metrics are weekly rollups, where weeks start on Monday and end on Sunday. Rollups are medians-of-medians, where we first take each individual active user’s median value for the week, then take the median of those values.

Calculating pre-Graphite median: For each active Graphite user in the selected time frame, we look at the median of each user’s 28 days of usage on GitHub before they ever created a stack with Graphite (in the selected repos). Then take the median again across all users. The goal is to create a fair benchmark of performance before these users starting using Graphite

Minimum 5+ active Graphite users: We won’t attempt to draw comparisons between developers using Graphite or GitHub if there are fewer than 5 active Graphite users in the selected timeframe.

Author stats

  • PRs merged: Total number of PRs where the user was an author

  • Time waiting on reviews: Median time of PRs authored by the user (that have open requests for reviews) were sitting without any reviews

  • Wait time to first review: Median time of a PR being marked as ready for review for the first time, and someone reviewing it

  • Publish to merge time: Median time between a PR being marked as ready for review for the first time, and it being merged

  • Review cycles until merge:

    • Computes the max number of review cycles any reviewer had until the PR was merged

    • Then across all PRs, computes the median of this number

Reviewer side stats

  • PRs reviewed: Total number of PRs that the user left a review on

Response time to review requests: Median of the difference in time between the user being requested (or re-requested) for a review, (referred to as review request/review pairs) for the given user

Insights example

Here’s an example of a sequence of events and how that activity will impact Insights stats:

  1. User X authors a PR in draft state that adds 10 lines of code, and deletes 5 lines of code

  2. User X publishes the PR, making it ready for review

  3. User X requests reviews from User A, User B, and User C

  4. User A requests changes on the PR

  5. User B leaves a review comment on the PR

  6. User X makes updates, and re-requests reviews from User A and User B

  7. User A leaves a review comment on the PR

  8. User B leaves a review comment on the PR

  9. User X makes updates, and re-requests review from User A

  10. User A approves the PR

  11. User X merges the PR

These events will affect each stats in the following ways:

  • PRs merged will go up by 1 for User X

  • PRs reviewed will go up by 1 for User A and User B (User C did not review, so no change there)

  • Response time to review requests will be computed for each reviewer on each review request/review pair as (*review time *- review request time). This will then feed into the reviewer’s aggregated median statistic.

    • User A response time for first review request = time at event #4 - time at event #3

    • User B response time for first review request = time at event #5 - time at event #3

    • User A response time for second review request = time at event #7 - time at event #6

    • User B response time for second review request = time at event #8 - time at event #6

    • User A response time for third review request = time at event #10 - time at event #9

  • Time waiting on reviews will be computed based on how long the author was waiting on reviews for that PR. This then feeds into the author’s aggregated median statistic.

    • Time waiting on reviews on the PR above for User X as author = (time at event #4 - time at event #3) + (time at event #7 - time at event #6)

    • Note that we didn’t count the time between event #10 and event #9 here because there was a review from User B on the PR—so the author wasn’t waiting on reviews during this time period. In case User X re-requested review from both User A and User B in event #9, we would also add (time at event #10 - time at event #9) into the computation above.

  • Wait time to first review will be computed based on how long the author waited until anyone reviewed their PR after it was published/ready for review. This will then feed into the author’s aggregated median statistic.

    • Wait time until first review on above PR for User X as author = time at event #4 - time at event #2
  • Publish to merge time will be computed based on how long it took for the PR to be merged after it was published/ready for review. This will then feed into the author’s aggregated median statistic.

    • Publish to merge time on above PR for User X as author = time at event #11 - time at event #2
  • Review cycles until merge for a PR will be computed as the max number of review request/review pairs that a reviewer had until the PR was merged. This will then feed into the author’s aggregated median statistic.

    • Review cycles for User A as reviewer = [(event #10, event #9), (event #7, event #6), (event #4, event #3)] (count = 3)

    • Review cycles for User B as reviewer = [(event #8, event #6), (event #5, event #3)] (count = 2)

    • Max review cycles on above PR for User X as author = max(3, 2) ⇒ 3

  • Lines of code added will go up by 10 for User X

  • Lines of code deleted will go up by 5 for User X