Are incident reviews about learning or tracking actions?
This question has sparked recent debate in incident management circles, including in my recent panel at SEV0 and in this post by Airbnb engineer Lorin Hochstein. Should the goal of an incident review be learning, or should it focus on tracking actionable improvements? When is the right time to discuss actions, and are they picked up just to make us feel better?
From my experience, learning from incidents and identifying actions are inseparable. Trying to split them into separate discussions or meetings—ostensibly to “focus on learning”—often does more harm than good.
To explore this in more depth, let’s break down what a balanced post-incident review should look like.
Here’s my baseline for running a healthy post-incident review. While the specifics will vary depending on the organization’s size, culture, and the context of the incident, this is—in my experience—a framework that works well for most teams.
Incident reviews are the meetings held after the initial issue has been mitigated, often within a few days. They’re typically attended by people directly involved in the incident, along with subject matter experts and relevant stakeholders. These meetings are usually cross-functional, with representation from engineering, customer support, and sometimes leadership.
In my experience, they generally shouldn’t involve people who weren’t involved in the incident, and who don’t have anything to contribute. The larger the “audience” in this meeting, the harder it is for folks to talk openly about their experience. It’s stressful enough walking through something that’s gone wrong without also feeling like you’re presenting to half of your engineering organization.
The objective of an incident review is straightforward: to align the group on what happened, ensure everyone understands things a little better, and identify areas for improvement. Learning and action discussions shouldn’t be seen as competing priorities but rather as complementary parts of the same conversation.
When I talk about discussing action items, I don’t mean diving deep into planning every minute detail. Nor do I mean carving out a specific time to talk about actions specifically.
I mean exploring improvements at contextually relevant points of the review process—improvements that might include preventing reoccurrence, enhancing reliability, or reducing future impact.
Some examples of how these conversations might arise:
Statements like this naturally arise as you discuss the incident, and they provide fertile ground for further learning and refinement. This type of “action item” discussion doesn’t derail the review; it enhances it.
In many cases, action items noted during the review actually spark further questions that deepen the learning experience. Consider the following questions:
By discussing action items at this level of abstraction people are learning about the system and we’re suggesting improvements. Feels like a pretty positive outcome.
With my case above made, let’s look at some specific criticisms I hear of this approach:
And these ones, pulled directly from Lorin’s blog post…
I’d encourage you to also read Why I don’t like discussing action items in incident reviews by Lorin Hochstein. It’s a different take on this subject, but I think there’s a reasonable overlap in philosophy.
In reality, a successful incident review may have no action items—or it may have a dozen. If you’ve made the right changes before the review, there’s no need to artificially invent actions. If new insights emerge during the review, note them down.
The goal is to learn and improve—whether that’s through tracking specific actions or simply building a shared understanding and deeper knowledge of what happened.
For me, I’ve never thought of it as a question of “learning or actions.” It’s about providing an environment where the most fruitful discussion happens. Learning can happen through action, and good actions are founded on better understanding of the systems we operate in.
I'm one of the co-founders, and the Chief Product Officer here at incident.io.
We were curious: where does the time go after an incident is resolved? To find out, we analyzed the post-incident process of 13,000 incidents and 14,000 follow-ups action items.
Post-mortem meetings can play a crucial role in fostering an environment of continuous learning. Here's how we do them!
With Post-incident flows, you can create a checklist of tasks and actions to ensure consistency across your teams and facilitate deeper learning.
Ready for modern incident management? Book a call with one our of our experts today.