GitHub ESSP
GitHub ESSP
Engineering System
Success Playbook
Contents
Introduction: GitHub
engineering zones and metrics
At GitHub, we know that better business outcomes aren’t driven just by good-quality code,
speed, or developer happiness in isolation. It’s actually when quality, velocity, and developer
happiness are working in unison that organizations see their best results. If you’re looking
for engineering to provide greater value to your business, it’s crucial to strengthen these —
let’s call them — foundational zones, and create better conditions for your teams to thrive.
This is the crux of GitHub’s Engineering System Success Playbook (ESSP) — a three-
step process that can help you drive meaningful, measurable improvements in your
organization, whether you’re looking to adopt a new AI tool like GitHub Copilot or identify
and unlock bottlenecks that have been hindering performance.
Inspired by multiple frameworks, including SPACE and DevEx, DX Core 4, and DORA, our
playbook offers a balanced and comprehensive approach, helping you assign metrics to
each “zone” that you can track over time and iterate as needed.
At the heart of our ESSP is a systems thinking1 approach that prioritizes long-term,
sustainable improvements. While quick wins can be a great way to get an initiative started,
they can produce negative downstream effects. For example, accelerating code review
turnaround time can speed up development, but without addressing the broader system –
like testing infrastructure and documentation practices – you may risk creating bottlenecks
downstream and compromising code quality.
This project was created in response to many customer requests for prescriptive guidance
on creating meaningful downstream impact from changes in their engineering systems—
1: A system is a group of interrelated, or interdependent parts that together serve a function or purpose (‘Thinking in Systems’ by Donalla Meadows). Systems thinking
brings a focus to the relationship between the multiple parts in the system (The Systems Thinker), recognising that the whole has emergent properties that are different to
the sum of its parts.
C O N T I N U ED O N N E X T PAG E PAG E — 3
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
often with the introduction of GitHub Copilot. We also engaged with DevEx and DevOps
metrics vendors to understand both the challenges and successes they’ve experienced
while helping customers elevate engineering performance or to justify the investment
in GenerativeAI. So these steps were created to balance the inherent complexity of
engineering success with practical, achievable steps for teams, including those earlier in
their improvement journey.
In this playbook, we’ll outline suggested metrics to monitor as part of your improvement
efforts for each zone. Keep in mind that these metrics are downstream, or lagging
metrics, and in the majority of cases should be complemented with leading metrics.
Both leading and lagging metrics may be measured using telemetry and/or survey data,
depending on your context, and the way these metrics are calculated will depend on
your teams’ engineering workflows and the systems supporting them—for example, you
may use Jira or ServiceNow alongside GitHub.
As you dig into this playbook, we encourage you to keep a few concepts in mind:
• Always bring a team perspective to improvement
• Select and use metrics with care to avoid gamification
• Balance the cost of measurement with the benefits of measurement
• Focus on improvements over time rather than overindexing on benchmarks
Engineering teams have the potential to fuel incredible change and accelerate business
outcomes. With GitHub’s ESSP, you can unlock engineering’s potential through creating
a culture of excellence that inspires and supports engineers to do their best work.
C O N T I N U ED O N N E X T PAG E PAG E — 4
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 5
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
The ESSP builds upon SPACE by identifying 12 specific metrics that help teams
improve engineering system performance. While the ESSP organizes these metrics
along developer happiness, quality, velocity, and business outcomes, they map
directly to the SPACE framework’s holistic approach.
C O N T I N U ED O N N E X T PAG E PAG E — 6
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 7
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 8
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Link to business outcomes: Enhanced code security and maintainability may reduce
risks, lower costs, and support ongoing innovation.
• Calculation advice: What event signals code vulnerability and exposure threat? What
quality attribute signals code maintainability? What quality attribute signals code
adaptability and reusability?
• Need to know: Availability of telemetry or survey data to evaluate both code maintainability
and security
• Tips: This metric could be calculated through a combination of analytics from GitHub
Advanced Security, SonarQube, or similar products or based on survey data. 4
Velocity
(Median) Lead time Deployment frequency
• SPACE dimension: Efficiency and flow • SPACE dimension: Activity
• Definition: The amount of time it takes • Definition: How often are releases
a commit to get into production 5 deployed to production 6
• Improvement (positive) direction: • Improvement (positive) direction:
Decrease is good Increase is good
• Link to business outcomes: Shorter • Link to business outcomes: Higher
lead times may enable faster deployment frequency may enable
responses to market demands. rapid innovation and faster customer
• Calculation advice: What event signals feedback cycles.
first commit for a release? What • Calculation advice: What event signals
event signals successful production successful production deployment
deployment (consider incremental (consider incremental release handling)
release handling)?
4 Survey questions can support answering this question where telemetry is not available, for example: It’s easy for me to understand and modify the code that I work with.
1 = Never; 2 = Rarely; 3 = Sometimes; 4 = Very Often; 5 = Always (This question is from DX’s Developer Experience Index (used with permission))
5 From DORA: https://dora.dev/quickcheck/
6 From DORA: https://dora.dev/quickcheck/
C O N T I N U ED O N N E X T PAG E PAG E — 9
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Link to business outcomes: Higher PR merge rates per developer may indicate effective
collaboration and accelerated delivery.
• Calculation advice: How many developers to include in calculation? What event signals
that the PR is merged?
• Tips: Focus on total PRs rather than calculating average for an individual and then
calculating the mean. GitHub recommends taking particular care in the calculation of this
metric. It should not be used to compare engineers to one another. Instead, the metric’s
purpose is to provide a measure of output adjusted for the number of engineers working
within a team or organization.
Developer happiness
(Median) Flow state experience (Median) Engineering tooling satisfaction
• SPACE dimension: Efficiency and flow • SPACE dimension: Satisfaction
• Definition: I have significant time for and well-being
deep, focused work during my work days. • Definition: How would you rate your
1 = Never; 2 = Rarely; 3 = Sometimes; overall satisfaction with the engineering
4 = Very Often; 5 = Always 7 tooling you use? 1 = Very unsatisfied,
• Improvement (positive) direction: 2 = Unsatisfied, 3 = Neutral, 4 = Satisfied,
Increase is good 5 = Very satisfied 8
• Link to business outcomes: Improved • Improvement (positive) direction:
flow state experience may enable Increase is good
engineers to deliver same or higher- • Link to business outcomes: Greater
quality work faster, with fewer errors and satisfaction with engineering tooling
interruptions. may reduce friction, enabling faster and
• Calculation advice: Ordered survey higher-quality software delivery.
responses for organization or team. • Calculation advice: Ordered survey
Identify middle value in results. Learn responses for organization or team.
more about developer flow. Identify middle value in results.
7 This question is from DX’s Developer Experience Index (used with permission)
8 This question is from DX’s Developer Experience Index (used with permission)
C O N T I N U ED O N N E X T PAG E PAG E — 1 0
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Business outcome
(Percentage) AI leverage (Percentage) Engineering expenses
to revenue
• SPACE dimension: Activity
• Definition: Opportunity being realized • SPACE dimension: Performance
due to effective engagement with AI, • Definition: The total engineering spending
through calculating the difference as a proportion of an organization’s total
between potential and current AI- revenue.
driven productivity gains across • Improvement (positive) direction:
employees working in engineering. Decrease is good
• Improvement (positive) direction: • Link to business outcomes: Lower
Increase is good engineering expense ratios may indicate
• Link to business outcomes: Higher efficient engineering investment and
AI leverage may reduce manual increased profitability.
engineering effort, or accelerate or • Calculation advice: What expenses are
enhance the quality of delivery with considered ‘total engineering’? What
increased cost efficiency. constitutes organizational revenue?
• Calculation advice: Average time- • Tip: Best monitored at organizational-level
savings associated with AI use. Average rather than team-level.
staff salary per week. Total staff who
could benefit from AI in engineering.
Total staff currently ‘engaged’ with AI for
engineering. Cost of AI per week
C O N T I N U ED O N N E X T PAG E PAG E — 11
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 1 2
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Three steps to
engineering success
These three steps are the heart of GitHub’s ESSP, as they highlight your current friction
points and manage your expectations for how changes will drive improvements. As
you consider your future state, GitHub recommends thinking across the zones: quality,
velocity, developer happiness, and how together they contribute to business outcomes.
As part of the three-step process, GitHub also recommends the use of leading
indicators—like close to code telemetry such as number of commits, and surveys—to
monitor the early impact of the agreed changes on your engineering system. Your choice
of leading indicators will depend on the friction points being addressed.
C O N T I N U ED O N N E X T PAG E PAG E — 1 3
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
1.1 Audit current processes, gather data, and understand organizational priorities:
C O N T I N U ED O N N E X T PAG E PAG E — 14
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Understand zone priorities: Engage with stakeholders to clarify which zones are
currently most critical for the organization given your business goals. Remember that
developer happiness is just as critical as the other zones. This helps to align the team’s
efforts with business goals and strategy.
• Map findings to the zones: Categorize each identified barrier by which zones it impacts
and onto your developer lifecycles.
• Prioritize the metrics to target: Once barriers are identified, prioritize which metrics
should be targeted for improvement. Consider any trade-offs between elements of
your desired future state and the barriers that are most actionable, keeping in mind
your business goals.
C O N T I N U ED O N N E X T PAG E PAG E — 1 5
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 1 6
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 17
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Identify potential solutions: Based on the barriers identified in step one, begin by
brainstorming possible changes that would reduce each barrier. For example, if a barrier
relates to slow deployments due to manual processes, one intervention might be
implementing automated deployment pipelines. If developer happiness is low, consider
initiatives that address workload balance or provide better tooling.
• Estimate cost and/or resource requirements: For each intervention, estimate the
resources required, including time, personnel, tooling, and budget. Consider both the
initial implementation effort and ongoing effort. Use this to support evaluation of the
feasibility of each intervention.
• Identify risks: Each change will have risks. For instance, automating a process may
inadvertently introduce new errors or bugs if not tested thoroughly. For cultural changes,
risks might include pushback from the team or slow adoption. Assess the potential
risks for each change, including both technical risks and people-related risks.
• Weigh the benefits against the risks and costs: For each change, clearly outline the
expected benefits and how they will support achievement of the future state. Make sure
to balance this with any potential negative impact on other areas of the business (e.g.,
increasing velocity at the cost of quality or developer happiness). Also account for the
C O N T I N U ED O N N E X T PAG E PAG E — 1 8
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Review with teams: Share the proposed changes with engineering teams to get
feedback. Are the changes realistic? Will the changes support long-term goals,
or are there concerns about their implementation? Developers, testers, product
managers, and other team members will have unique insights into the practicalities of
implementing changes. If you have undertaken a pilot, share the findings from the pilot.
• Secure buy-in: For more significant changes, secure buy-in from leadership and other
stakeholders. Present the expected benefits alongside the potential risks and cost or
resource requirements. It’s important that there is alignment across all levels of the
organization, especially when the interventions involve process changes or resource
investments. Also be realistic about the timeframe for implementation and the
realization of benefits.
• Incorporate feedback: Be open to adjusting interventions based on feedback from
stakeholders, including those involved in any pilots. Some changes may need to be
deprioritized if they are deemed too risky or resource-intensive, while others may be
refined based on team input.
C O N T I N U ED O N N E X T PAG E PAG E — 1 9
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 2 0
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 2 1
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
reduce deployment frequency metrics in our velocity zone. We weighed the impact to our velocity
and determined how much of an impact we were able to withstand in order to see gains in quality
metrics. To counteract some of the reductions in velocity, the team also proposed increasing the
number of changes that could be deployed at once. This analysis was done very carefully to make
sure that an increase in changes would not result in a reduction of quality.
Once we decided which process and tooling changes to pursue—just like our features—we then
took an incremental approach to roll-out. You can work in incremental changes in two different
ways:
• The number of changes you make at a given time. Keep the scope small
so you know which change is driving what impact
• In terms of the distribution of the change (then scale to build confidence)
We also used a test application that allowed us to A/B test our process and tooling changes to
more accurately understand their impact on key metrics.
C O N T I N U ED O N N E X T PAG E PAG E — 2 2
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Track key metrics: Once changes are implemented, track the identified metrics across
the zones. Compare the new metrics with the baseline established in step one to
evaluate the impact of the intervention. However, be realistic about the time it takes
for metrics to shift and expect some variance in performance rather than consistent
gains. Most change initiatives will require the use of a set of leading indicators. Often,
qualitative data like surveys are a useful leading indicator in addition to close to code
metrics such as pull request review times, depending on the current and future state
and barriers being addressed by the changes. Learnings from any pilots can be useful in
understanding likely timeframes to achieve downstream improvements.
• Gather qualitative feedback: In addition to metrics, gather feedback from developers,
operations, and other stakeholders on how the changes are impacting their day-to-day
work. Use interviews and team retrospectives to understand whether the changes are
positively affecting team morale, collaboration, or overall satisfaction.
• Identify early wins and challenges: Keep an eye out for both early successes and
challenges. Celebrate small wins, such as reductions in pull request review times
or improved test coverage, to build momentum. On the flip side, be prepared to
identify and address any resistance or unforeseen issues early, before they grow
into larger problems.
C O N T I N U ED O N N E X T PAG E PAG E — 2 3
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
• Analyze what’s working and what’s not: After an initial period of implementation, review
the data and feedback to confirm that the changes are having the desired effect. Are
the target metrics improving? Are there trade-offs that need to be reconsidered? For
example, are quality targets being maintained while velocity is improving? It’s essential
to critically assess whether the changes are solving the barriers identified in step one.
• Pivot if necessary: If changes are not delivering the expected results, don’t hesitate to
pivot. It’s better to adjust mid-course than to persist with solutions that aren’t working.
Revisit the other potential actions from step two, and consider alternative approaches
or adjustments.
• Maintain continuous feedback loops: Make monitoring and feedback an ongoing
process. Don’t treat implementation as a one-time effort. Use team retrospectives,
stakeholder reviews, and performance dashboards to maintain a cycle of continuous
improvement. Regularly check in on the health of the zones and be proactive in
adjusting the changes as needed. Consider using automated alerting to make sure
that if a metric is falling outside expected performance ranges, it can be reviewed and
acted upon.
C O N T I N U ED O N N E X T PAG E PAG E — 24
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
impact. Surveys are a great tool for the earlier stages of an intervention. Remember, it
may take time for the team to adjust and for the changes to be fully embedded.
• Keep iterating on the changes: Remember that even after successful implementation,
further improvements can always be made. Teams should be encouraged to treat
the process as ongoing and remain open to refining changes as new challenges arise.
Changes in operating circumstances can also prompt the need to consider further
iterations.
• Watch out for unintended consequences: Some changes may introduce new friction
points or affect other areas of the workflow in unexpected ways. For example, speeding
up deployments may lead to more frequent post-release bugs if the quality zone isn’t
balanced. Be vigilant in identifying these side effects and address them promptly.
• Check in on psychological safety: Make sure that teams still feel comfortable speaking
up about issues post-implementation. Teams should feel empowered to offer honest
feedback about what’s working and what isn’t, without fear of judgment.
• Evaluate long-term impact: Over time, make sure that the improvements are sustained
and that new challenges aren’t introduced. Look for enduring improvements in team
performance and morale.
• Use feedback for further learning: Treat failures as opportunities for learning. If a
change doesn’t work, use the data and feedback gathered to understand why, and
apply those
C O N T I N U ED O N N E X T PAG E PAG E — 2 5
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 2 6
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 2 7
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Tailoring
The GitHub ESSP should be tailored to align with your team’s specific needs, workflows,
and tooling. Rather than applying a universal approach, tailoring enables teams to select
metrics that directly reflect their goals, context, and budget. For instance, some teams may
choose to focus more heavily on developer happiness if morale is hindering engineering
system performance, while others might prioritize velocity to meet business goals.
Tailoring also involves deciding how to measure metrics—opting for telemetry data when
automated tracking is practical, or developer surveys when more nuanced feedback
is needed. Measurement is a tool—valuable when it supports improvement efforts
but not an end in itself. It’s essential to invest only as much as needed to facilitate or
evidence meaningful progress, avoiding the temptation to over-engineer the engineering
measurement system.
Additionally, tailoring includes complementing downstream metrics (e.g., deployment
frequency) with leading indicators that offer early signals of friction or improvement. By
tailoring such elements, the ESSP can reflect the diverse ways that engineering teams
work. As you tailor and select metrics, remember to collect data across the ESSP zones
C O N T I N U ED O N N E X T PAG E PAG E — 2 8
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
(developer happiness, quality, velocity, and business outcomes) and include at least
three of the SPACE dimensions (satisfaction, performance, activity, communication and
collaboration, and efficiency and flow.)
Change management
Change management is essential for achieving success within the GitHub ESSP, ensuring
that teams can adopt and sustainably use new metrics, tools, and practices effectively.
Frameworks like ADKAR (Awareness, Desire, Knowledge, Ability, Reinforcement), Kotter’s
eight-step change model, or the three-stage change model associated with Lewin, provide
structured approaches that can be tailored to guide engineering teams through change.
For example, ADKAR’s focus on awareness and desire is useful when introducing new
metrics like those in the playbook, helping teams understand the role of these metrics
in guiding sustainable improvement. Kotter’s emphasis on building a coalition can rally
support across teams, especially when adopting telemetry or new measurement methods.
By applying these frameworks, change management can help teams feel prepared and
supported.
Growth mindset
GitHub’s ESSP is most powerful when approached with a growth mindset that values
learning as success in itself. This means recognizing that not every intervention will work as
intended on the first try—and that’s okay. Each attempt, whether it leads to immediate gains
or requires recalibration, offers valuable insights that propel teams forward. By embracing
the idea that failure is part of the process, teams can take bolder steps in understanding
bottlenecks, experimenting with solutions, and refining their practices. This mindset
fosters resilience, allowing teams to adapt, learn, and ultimately build an engineering
culture where each iteration brings them closer to sustainable improvement.
Gamification
When thoughtfully designed, gamification can foster a positive and motivated engineering
culture. Drawing from behavioral economics, gamification is more effective when it
aligns with intrinsic motivators—such as the satisfaction of mastering a skill, solving
complex problems, or contributing to team goals. For instance, rewarding developers with
C O N T I N U ED O N N E X T PAG E PAG E — 2 9
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
However, it’s also necessary to be aware of the risks of incentive misalignment, where
rewards can encourage undesirable behaviors. For instance, using leaderboards to drive
rapid code reviews may prompt rushed reviews that compromise code quality. Similarly,
tracking and rewarding individual code contributions can lead to a focus on quantity over
quality, potentially increasing technical debt. It’s essential to recognize both subtle and
overt incentives that the monitoring of metrics can introduce.
C O N T I N U ED O N N E X T PAG E PAG E — 3 0
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 3 1
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
dimensions of SPACE (see Page 5), organizations can ensure they account for the
complexity of engineering work at the individual, team, and system levels.
This approach helps avoid overly simplistic productivity measures and instead fosters a
nuanced understanding of the factors that drive sustainable performance, collaboration,
and satisfaction. By leveraging SPACE, organizations can shape initiatives that align with
both developer well-being and business impact, ensuring an evidence-based path to
engineering success.
C O N T I N U ED O N N E X T PAG E PAG E — 3 2
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
C O N T I N U ED O N N E X T PAG E PAG E — 3 3
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Engineering
Description Teams wait too long to release, deploying large
batches of code at once.
antipatterns
Lack of CI/CD pipeline maturity.
Preference for ‘all-at-once’ (or quality) certainty.
Strict compliance requirements.
Long review cycles between PR and deployment.
C O N T I N U ED O N N E X T PAG E PAG E — 3 4
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Description Developers spend too much Building overly complex solutions Ignoring or deferring technical debt,
time perfecting code or adding for simple problems. allowing inefficient and vulnerable
unnecessary features. systems to persist.
Quality impact Increased complexity introduces Complex systems are more prone Code becomes brittle and bug-
more potential for bugs without to bugs and harder to maintain. prone, leading to poor system
added value to user. health.
Velocity impact Adds unnecessary time to Slows development as complexity Increases time to develop new
development as teams over-focus adds overhead to build and features as workarounds grow.
on perfection. maintain systems.
How AI could help Use GitHub Copilot to simplify Use GitHub Copilot to refactor Use GitHub Copilot to create
code and remove redundant code. existing code. This could be to tests and refactor existing code.
make the code more modular, or to This could be to make the code
suggest a simpler way of solving the more modular, or to suggest a
problem. simpler way of solving the problem.
Autofix may reduce effort and
increase satisfaction with starter
suggestions in PRs.
Friction requiring non-AI Product management decisions Overdesigning systems to solve Prioritize and allocate engineers to
intervention about feature prioritization. edge cases that rarely occur. address the technical debt.
Potential leading or Work in Progress ↑ Developer satisfaction with delivery Code complexity ↑
additional metrics or cadence ↓
indicators that may Late-in cycle code churn ↑ Large blocks of commented out
indicate this antipattern Usage of features/sub-features ↓ code ↑
Usage of features/sub-features ↓
[↑ ↓ trend suggestions Cognitive complexity ↑ Duplicated Blocks ↑
antipattern] Hardcoded values and secrets ↑
Dependency issues ↑
Zone metrics that may Lead time ↑ Code security and maintainability ↓ Code security and maintainability ↓
indicate this antipattern
Lead time ↑ Lead time ↑
[↑ ↓ trend suggestions
antipattern] Change Failure Rate ↑
C O N T I N U ED O N N E X T PAG E PAG E — 3 5
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Description Teams receive vague or incomplete Relying on manual steps for Relying on manual testing or
requirements, leading to deployment instead of automation. insufficient test automation,
misunderstandings. causing delays.
Potential root-causes Pressure to start development Perception that manual is ‘good Belief in thoroughness of manual
quickly. enough.’ testing.
Immature product discovery Fear of effort needed for Resource constraints for
process. automation. automation.
Frequent priority shifts from Lack of investment in DevOps Limited familiarity with modern
stakeholders. practices. test tools.
Previous experience with brittle,
costly, or flaky tests.
Quality impact Poorly defined requirements lead to Manual deployments introduce Lack of thorough testing introduces
incorrect or low-quality features. inconsistent outcomes that can more bugs into production.
lead to post-deployment bugs.
Velocity impact Time wasted clarifying Slows releases. Delays releases as testing takes
requirements or building incorrect longer.
features.
How AI could help Stay tuned: GitHub’s AI powered Use GitHub Copilot to create Use GitHub Copilot to create test
platform continues to evolve automation, such as GitHub Action suites, and automate CI workflows,
workflows, to replace manual to remove frictions.
deployments.
Use GitHub Copilot to troubleshoot
why a deployment automation has
failed.
Friction requiring non-AI Engaging with stakeholders to Inconsistent processes and The need for a robust
intervention ensure real-world needs are human reluctance to adopt testing strategy aligned with
reflected in the requirements. automated deployment pipelines. the project’s goals
Potential leading or Time spent in meetings ↑ Count of manual steps per Automated test coverage ↓
additional metrics or deployment ↑
indicators that may Work in Progress ↑ Time spent on manual testing ↑
indicate this antipattern Dwell (delay) time during CI/CD ↑
Rework ↑
[↑ ↓ trend suggestions Deployment duration ↑
Developer frustration ↑
antipattern]
Zone metrics that may Flow state experience ↓ Deployment frequency ↓ Change failure rate ↑
indicate this antipattern
Lead time ↑ Failed deployment recovery time ↑ Deployment frequency ↓
[↑ ↓ trend suggestions
antipattern] PRs merged per developer ↓ Change failure rate ↑ (Median) Lead time ↑
Engineering tooling satisfaction ↓ Engineering tooling satisfaction ↓
C O N T I N U ED O N N E X T PAG E PAG E — 3 6
G I T H U B ’ S EN G INEER IN G SYS T EM S U C C ES S PL AY B O O K M AY 2 0 25
Description Teams operate in silos, failing to Feedback from testing, users, or Constant addition of features or
share data, tools, or processes other stakeholders is not provided changes mid-development without
across teams. in a timely or consistent manner. proper evaluation.
Potential root-causes Incentives misaligned across Waterfall mindset undervaluing Unclear project boundaries.
teams. iteration.
Poor change management
Culture prioritizes team-specific Feedback viewed as an end-phase practices.
goals. activity.
Culture discourages saying “no” to
Historical habit of independent Lack of real-time feedback tools. requests.
operation.
Quality impact Inconsistent processes and tools Bugs and user issues linger due to Rushed development due to scope
result in lower-quality handoffs delayed feedback. creep often leads to more bugs and
between teams. lower quality.
Velocity impact Cross-team dependencies lead to Slows iteration cycles, as engineers Introduces unplanned work that
delays when teams aren’t aligned. aren’t able to adapt quickly. delays original timelines.
How AI could help Copilot features can help improve Use GitHub Copilot for Pull Stay tuned: GitHub’s AI powered
documentation and code Requests to automatically analyze platform continues to evolve
explanations. pull requests and suggest changes
to provide a more consistent
feedback loop.
Developers can use GitHub
Copilot to ask questions about a
pull request, providing for a more
informed pull request review
that leads to a more consistent
feedback loop
Friction requiring non-AI Cultural issues or lack of Human communication and Managing stakeholder expectations
intervention communication between teams. prioritization of feedback. and ensuring a disciplined
approach to scope management.
Potential leading or Cross-team collaboration Feedback frequency ↓ Scope changes per sprint ↑
additional metrics or frequency ↓
indicators that may Feedback quality ↓ Ratio of issue types per sprint ↑
indicate this antipattern Handoff delays ↑
Customer satisfaction ↓ Time spent on unplanned work ↑
[↑ ↓ trend suggestions Rework frequency ↑
Age of PR’s last human activity ↑
antipattern] Poor meeting attendance ↑
Zone metrics that may Lead time ↑ Lead time ↑ Lead time ↑
indicate this antipattern
PRs merged per developer ↓ Flow state experience ↓
[↑ ↓ trend suggestions
antipattern] Deployment frequency ↓
PAG E — 3 7
W R I T T EN BY G I T H U B W I T H