Out of a hundred people, only five are managing to get the insights they need from the myriad monitoring tools they use to track the performance, availability, and security of their digital assets.
The other ninety-five? They might as well as be staring at cute but ineffective Venn diagrams.
One of the causes of missing insights is not a dearth of tools or even a lack of data; it’s the continuous bias in the data.
Data bias is caused by opinionated curation within a system. Opinionated curation is the result of decisions made regarding everything from what data is collected and from what systems, to how visualizations are displayed on dashboards.
One of the most significant sources of data bias is the legacy of agent-based systems, which require additional software to be deployed on any system from which you want to collect data. Agents are usually paired with an analytics platform, and in the past the pricing of these products often included the cost of agents. Discounted for large deployments, of course, but nonetheless a costly proposition.
The cost of deploying—and managing—agents introduced data bias because it limited the systems from which you could collect data. Opinions—experienced or not—regarding the worthiness of monitoring a given application or system often drove the deployment decision.
The volume of data generated, too, is a source of continuous bias because it led to decisions based on the opinion of the value of specific data points. Do you really need all those metrics, or can we just limit it to three or four? The opinionated curation of metrics means there’s no way to validate shifts or changes in the data overall. Those shifts or changes could be indicative of a potential problem—or risk—but are missed because the data was deemed insignificant.
Finally, dashboard decisions about visualization further bias interpretation and are often based on skills and experiences that may not be shared with others who use the dashboard. Even the choice of graph can introduce bias. This is particularly true for time-series based operational metrics like performance and uptime.
Bar charts are often used to chart time-series data but aren’t as impactful as line charts. A bar chart forces us to compare bar heights to understand shifts in response time rather than allowing the shape of the line show us what’s happening. These simple decisions can have a profound impact on operators who rely on visualization to offer insight into the operating state of a system.
All these decisions introduce bias, continually, into the data and impact our ability to interpret it—and thus gain insight into what it really means.
If we want better decisions in a digital as default world, we’re going to need better data and that means eliminating as much data bias as we can.
This is one of the reasons OpenTelemetry is so promising. Standardizing the way telemetry is generated and ingested using open-source (and thus lower-to-no-cost) agents will go a long way toward eliminating one of the root causes of data bias: the IT budget. By ensuring you can collect telemetry from every system rather than just a few deemed "worthy," you eliminate a significant source of bias in your data. That’s why our vision of edge includes the notion of embedding telemetry generation into the platform itself, so it’s always available, everywhere.
Data lake as a service, too, can be an effective means of address the bias introduced by curation due to volume and cost to store over time. By outsourcing scale and capacity, organizations can more freely ingest more telemetry, making it easier to uncover anomalies and patterns that impact performance and more. Over the past five years we’ve seen a growing menu of such services, often rolled up as part of the XaaS category that continues to consume IT budget at an increasing rate. Deloitte noted in a report on XaaS this year that nearly half of organizations will allocate at least half of its IT spending to XaaS.
Finally, by migrating from visualizations to insights, the issue with bias in dashboard delivery of operational data can be addressed. Rather than just displaying snapshots of data points in time, insights deliver information based on patterns and relationships discovered in the data. What’s more, insights can eliminate unnecessary fire drills caused by reliance on binary metrics. Modern systems are built to fail and recover automatically. Being notified whenever that happens can cause alert fatigue that drives down productivity. The ability to analyze telemetry in the context of an entire user flow means understanding when a failed component requires intervention and when it doesn’t. Migrating from a mode that relies on interpreting multiple visualizations to leveraging comprehensive insights can eliminate some of the bias inherent in visualization.
As business approaches a default-digital model, it necessarily relies more and more on data. Eliminating bias in that data will be a critical step in ensuring that decisions based on that data will lead to the best outcome possible for customers and the business.