Observability Is Cultural
In order to leverage observability, we need a significant shift in our corporate culture that encapsulates the entire company and goes beyond the tools.
Join the DZone community and get the full member experience.
Join For FreeI’m guilty of applying the word "debugging" to practically anything. My kids' Legos won’t fit, let’s debug that. Observability is one of the few disciplines that actually warrant that moniker: it is debugging. But traditional debugging doesn’t really fit with observability practices. I usually call it “precognitive debugging." We need to have a rough idea in advance of what our debugging process will look like for effective observability troubleshooting.
Note that this doesn’t apply to developer observability, which is a special case. That’s a more dynamic process that more closely resembles a typical debugging session. This is about more traditional monitoring and observability: where we need to first instrument the system and add logs, metrics, etc. to cover the information we would need as we will later investigate the issue.
I wrote before about the scourge of overlogging. The same applies to observability metrics, as we collect more and more data the costs for retention and processing quickly outweigh the benefits of observability. We end up with a bigger problem altogether. We need to pick our battles, log the “right amount” and monitor the “right amount," no more and no less than we need. For that, we need to understand the risks that we’re dealing with and try to maximize overlap in our investigation.
Chaos Engineering as Inspiration
In the tradition of Chaos Engineering, we would organize a “game” orchestrated by the “master of disaster” to practice disaster readiness. This is a wonderful exercise and a great way to build that “muscle." It isn’t the right fit for an observability architecture, since observability deals with nuance as opposed to “fire."
Observability requires a similar game, but a deliberate one, where our team competes on finding the ways in which our system can fail. Think of it as Bingo. Once we have a spreadsheet full of potential failures, we need to map out the failures to the observability we would like to have for every potential failure. For example, in case of a hack, we’d like to have the user id logged when accessing any restricted resource.
Once we chart all of those desires, we can review them, and try to unify some metrics and logs. Then, we can implement them so our observability can answer everything we need to track down an issue.
Will we miss some things?
Obviously. That’s part of the process. We will need to iterate and tune this. It will probably require a reduction of volume for some expensive data points to keep the costs reasonable. We will undoubtedly run into issues that aren’t covered by observability (or whose observability coverage isn’t obvious). In both cases, we will need some help.
Do We Need Experts?
Some observability fans assume that we no longer need domain experience to debug a problem. Given a properly observable system, we should be able to understand the problem without knowing anything about the system.
While I agree that an expert in debugging can probably solve a problem faster, possibly faster than a domain expert, I still have my doubts. Over the course of a decade, I was a consultant and I would go to companies where I used profilers, debuggers, etc. As part of that job, I found the issues that escaped people who were greater domain experts than I was. So there’s some merit behind that claim.
But debugging requires some familiarity with the system that we’re trying to understand. It’s like diagnosing through Google. We might occasionally find the cause better than our GP, but probably no better than an expert. Obviously, there are exceptions to the rule, but in my experience, experience matters for any type of debugging.
A Dashboard of Our Own
One thing I see often is a universal “one size fits all” dashboard in a company. Grafana is a fantastic tool with remarkable flexibility, yet some expose its visualizations as a single company dashboard. There should be at least three dashboards for the application:
- High level: CTO/VP R&D level; this focuses on business metrics, users, reliability, costs
- DevOps: Low-level information about the environment
- Developers: Application-specific metrics and platform information
There’s a lot of overlap there, but we need custom dashboards. The whole idea of the dashboard is to see everything that matters in one place. CPU utilization on the container might be interesting to me in general, but more likely than not, it will just be a distraction. I want to know if there’s a problem with the authorization system because users are experiencing increased error rates logging in. These metrics should be front and center.
When I open a new tab in my browser, I see Grafana. This should be the home page for every team member. The “healthy” view of our system should be etched into our minds so we can instantly notice small deviations in the environment and act accordingly.
Growing With Observability
As our system grows, we need to include observability and metrics in the pull request that introduces a feature. Nothing can launch without observability on day one. It should be etched into the code review process and should be on par with test coverage requirements.
Unlike test coverage, we have no metric we can rely on to verify that observability is sufficient for the rapidly evolving needs; so at this time, this is a heavy load on the shoulders of the reviewers. But there’s an even bigger load: cost. As we grow, these changes can affect cost, which can suddenly spike to bankruptcy-inducing heights. Cost isn’t always easy to monitor, but it’s a gauge we should look at on a daily basis. By keeping track of that metric and catching spikes in cost early on, we can keep our systems stable and manageable without giving up cost-effectiveness.
Finally
Some engineers have an over-infatuation with metrics. I’m not one of them. Some things can’t be measured: the value of personal relationships, the value of a team, and a community. Because of this obsession, observability is gaining in popularity. That’s good and bad. With this obsession, we sometimes over log and observe, which results in poor performance and cost overruns.
We should apply observability with a scalpel, not with a shovel. This shouldn’t be something we delegate to the DevOps team as an afterthought. It should be a group effort that we constantly refine as we move along. We should keep our pulse on our metrics and have domain-specific dashboards to constantly keep the things that matter in our peripheral vision. Observability doesn’t matter if we don’t bother looking.
Published at DZone with permission of Shai Almog, DZone MVB. See the original article here.
Opinions expressed by DZone contributors are their own.
Comments