From Monitoring to Observability: A Fundamental Shift
Traditional infrastructure monitoring was predicated on a knowable system. You knew what your infrastructure consisted of — a defined set of servers, applications, and network equipment. You defined metrics for each component (CPU utilization, memory usage, request latency) and set thresholds that triggered alerts when those metrics exceeded acceptable values. Monitoring was essentially a surveillance system for a known environment.
Cloud-native architectures broke this model. A modern cloud application might consist of hundreds of microservices, deployed as containers that are created and destroyed dynamically by Kubernetes based on traffic patterns. The infrastructure is ephemeral — a service instance that exists for ten minutes does not lend itself to traditional monitoring dashboards. The interactions between services are complex, opaque, and highly variable. And the failure modes are often emergent properties of the distributed system rather than failures of individual components — a latency spike in one service cascades through dozens of dependent services in ways that are difficult to trace.
Observability emerged as the conceptual response to this challenge. Observability — a term borrowed from control systems theory — refers to the ability to understand the internal state of a system from its external outputs. An observable system is one where, when something unexpected happens, you can answer the question "what is happening and why?" without having to pre-define exactly what metrics to look at. Achieving observability requires a different approach to telemetry collection, storage, and analysis than traditional monitoring.
The Three Pillars: Metrics, Logs, and Traces
The observability community has converged on three types of telemetry data — often called the three pillars of observability — that together provide a comprehensive view of distributed system behavior:
Metrics are numerical measurements sampled at regular intervals: request rates, error rates, latency percentiles, resource utilization. Metrics are efficient to collect and store at scale, and they are well-suited to alerting on known conditions. Metrics answer the question "is something wrong?"
Logs are timestamped records of discrete events: a request arrived, a query was executed, an error was encountered. Logs provide rich context about specific events and are essential for debugging — but they are expensive to store at scale and require careful structuring to be query-able across the volumes that modern systems generate.
Traces are end-to-end records of individual requests as they flow through a distributed system, capturing the timing and outcome of each service call in the request's path. Traces are essential for understanding latency in distributed systems and for identifying which service in a complex dependency chain is causing a performance problem. Distributed tracing is the newest of the three pillars and the one where adoption is still most incomplete across the industry.
The holy grail of observability — which the leading platforms are increasingly achieving — is the correlation of these three data types: being able to jump from a metric anomaly to the relevant logs and traces that explain it, all in a single workflow. This correlation requires sophisticated data modeling and query capabilities that are still an area of active development in the observability market.
The OpenTelemetry Revolution
One of the most important developments in the observability market has been the emergence of OpenTelemetry as an open standard for telemetry instrumentation. Prior to OpenTelemetry, every observability vendor had its own proprietary instrumentation library. If you instrumented your application for Datadog, you were locked into Datadog — switching to a different backend required re-instrumenting your entire application. This vendor lock-in was a major barrier to observability adoption and gave incumbent vendors pricing power that was not always earned by product quality.
OpenTelemetry, now a Cloud Native Computing Foundation (CNCF) project with broad industry support, defines a vendor-neutral set of APIs, SDKs, and collector infrastructure for generating and exporting metrics, logs, and traces from applications. Applications instrumented with OpenTelemetry can export their telemetry to any compatible backend — or multiple backends simultaneously — without code changes. This decouples instrumentation from backend vendor choice and dramatically changes the competitive dynamics of the observability market.
For observability platform vendors, OpenTelemetry is both a challenge and an opportunity. It erodes the instrumentation lock-in that legacy vendors relied on, increasing competition. But it also dramatically lowers the instrumentation barrier for adopting new observability tools, expanding the total available market by making it easier for organizations to try new platforms. The winners in the OpenTelemetry world will be the platforms that provide the best analysis, visualization, and actionability on top of the commoditized telemetry layer.
Observability and Security: The Convergence Thesis
One of our highest-conviction investment themes is the convergence of observability and security. The data that powers observability — logs, metrics, traces, network flow data — is also the data that powers threat detection. An observability platform that can detect anomalous application behavior (a service making unusual external network calls, a user accessing an unusual volume of data, an authentication pattern that deviates from baseline) is also capable of detecting security incidents.
The convergence thesis suggests that the next generation of enterprise security tools will be built on observability infrastructure, and the next generation of observability platforms will incorporate security analytics as a first-class capability. This creates interesting investment opportunities at the intersection of the two markets — companies that can serve both the DevOps and security buyers with a unified data platform are potentially addressing a combined market that is substantially larger than either individually.
The practical challenge of this convergence is organizational: security teams and platform engineering teams have different buyers, different workflows, and different tool preferences. Companies pursuing the convergence thesis need to be thoughtful about how they navigate the organizational complexity of serving two distinct buyer personas with a single product.
The Cost Challenge in Observability
The rapid growth of observability data volumes has created a significant cost challenge for enterprises. A large cloud-native application generating metrics, logs, and traces at full fidelity can produce hundreds of gigabytes or even terabytes of telemetry data per day. Storing and querying this data at the retention periods enterprises require for compliance and incident investigation can cost millions of dollars annually.
Observability cost optimization has become a major concern for platform engineering teams, and it is driving demand for a new generation of telemetry management tools: platforms that intelligently sample, filter, and route telemetry data to reduce storage costs without sacrificing the visibility needed for incident investigation. This is a genuinely hard technical problem — determining which data to retain and which to discard requires understanding what data will be needed for future investigations, which is inherently unpredictable.
Key Takeaways
- Observability — understanding complex distributed systems from their outputs — is a foundational operational capability for cloud-native enterprises.
- The three pillars (metrics, logs, traces) together enable comprehensive system understanding; correlation across all three is the current frontier.
- OpenTelemetry is commoditizing instrumentation and shifting competition to analysis, visualization, and actionability.
- The convergence of observability and security is creating opportunities for platforms that serve both DevOps and security buyers.
- Observability cost optimization is a growing market as telemetry data volumes outpace traditional storage economics.
Conclusion
Observability is not a feature — it is the foundation on which reliable, secure, and well-operated cloud infrastructure is built. As cloud environments grow more complex, the ability to understand and debug those environments in real time becomes increasingly valuable. The companies building the next generation of cloud-native observability platforms are addressing one of the most fundamental needs in enterprise technology. At Key AI Ventures, observability is a key investment area within our cloud infrastructure thesis. Connect with our team or explore our portfolio to learn more.