Necessity is the mother of invention, yet it often sows seeds of complexity. As applications have evolved to meet user demand for more sophisticated, performant and personalized experiences, the complexity of their underlying architectures and componentry has intensified. Because of this, effective monitoring and debugging is crucial to any product’s success. However, achieving high fidelity often requires advanced tooling and expertise to properly instrument apps across distributed architectures, multi-cloud environments, third-party API integrations, embedded machine learning models and web/mobile/edge deployments among others.
Data consistency is a common observability challenge. Each component within these intricate systems will often emit its own unique telemetry format, leading to continuous streams of mixed signals for SREs (Site Reliability Engineers) to decode. Not only is data processing and correlation a challenge, but so is managing its compounded growth, which some analysts estimate is as high as ~23% YoY.* These complexities lead to escalating infrastructure, software and engineering resource costs, often without proportionally increasing insights.
Furthermore, the industry struggles to provide coverage across all phases and personas of the product life cycle. We believe the scope of observability is expanding, and no longer limited to classical definitions around systems and IT operations. This includes a ‘shift left’ to provide developers with faster and more effective troubleshooting, debugging and code-level monitoring, as well as a ‘shift right’ to provide business stakeholders visibility into customer usage and behaviors influencing product KPIs.
Finally, as the aperture of modern applications expands, the industry must continuously extend instrumentation to new environments and emerging workload types such as LLM-powered components.
These challenges are opening the door for the next generation of disruptors that promise greater coverage, interoperability and cost efficiency, without sacrificing depth or clarity of insight. In recent years, the industry has experienced a fresh wave of innovative new capabilities and approaches to observability, including data pipelines, in-stream analytics, database & storage layer optimizations, AI-powered root cause analysis, standardized data formats and more.
Sapphire Ventures has been fortunate enough to back some incredible companies across the observability ecosystem over the years. Working with each of these teams has given us a unique vantage point into the market and, as the observability ecosystem evolves, we’re excited to share our perspective on the key trends shaping this space in 2024 and beyond:
Observability Pipelines providing real-time filtering, enrichment, normalization and routing of telemetry data
Insertion of LLMs to simplify the user experience and enhance both analytics and downstream automation
Emergence of AI/ML Observability tools to help monitor and optimize AI/ML workloads
Industry-Wide Standardization and OpenTelemetry
Bring-Your-Own Storage Backend to enable more flexible and efficient utilization of infrastructure
Visibility into the CI Pipeline to optimize end-to-end software development processes
Linking Business Outcomes to Systems Data to correlate product-level data with backend performance
2024 Observability Trends
1. Observability Pipelines
As data volumes and costs continue to rise, organizations are requiring more granular control over their telemetry data, from collecting, processing and routing to storage and archival. One approach that's gained significant traction is observability (o11y) pipelines, which have emerged as a powerful way to lower overall ingestion costs and volumes.
Pipelines serve as a telemetry gateway and perform real-time filtering, enrichment, normalization and routing, often to cheap/deep storage options such as AWS S3, reducing dependencies on more expensive and proprietary indexers. In addition to cost savings, another common benefit we see, particularly in the enterprise, is data (re)formatting. In this case, pipeline customers convert legacy data structures to more standards-based formats on the fly (e.g., OpenTelemetry, see trend #4), without needing to ‘touch’ or reinstrument legacy code bases.
Data pipelines are packaged as both standalone solutions or as a native subcomponent of a broader observability suite. Regardless of starting position, we believe players in this space will continue expanding, supporting a wider range of operational use cases over time (e.g., Cribl has expanded beyond pipelines into app monitoring w/ AppScope).
2. Insertion of Large Language Models (LLMs)
The observability market has used traditional ML models for years to provide anomaly detection, capacity forecasts, outage detection and other predictive capabilities (often branded as AI-Ops solutions). More recently, platform providers have started using LLMs to inject additional intelligence into their offerings and simplify the overall user experience.
LLMs enable natural language to configure platforms, instrument applications, create dashboards and assemble queries, all of which are then translated into platform-specific commands and syntax. Once a baseline configuration is established, LLMs can also learn and adapt over time. They can personalize the management console to show frequently viewed app components and associated metrics based on user affinity and previous interactions.
LLMs are also making it easier to understand complex system alerts. They take alerts that are full of esoteric technical jargon and turn them into text that is easier for people to read. Because LLMs can ingest and analyze large volumes of unstructured data, companies like Flip.ai are finding new ways to apply them to more traditional log analytics and AI-Ops capabilities, identifying patterns, inferring meaning (e.g., sentiment analysis of user feedback) and determining root cause across complex systems.
Finally, as the predictability and explainability of LLMs improve over time, we expect them to start making decisions on their own. Agentic capabilities, or the capacity for decision-making, are likely to manifest, particularly in support of incident management workflows, where models will be granted autonomy to both assemble runbooks on the fly and take steps to execute potential downstream fixes.
It's important to note that while the application of LLMs to o11y is promising, their effectiveness in interpreting technical data and generating actionable insights, without significant domain customization and use of case-specific training, remains a significant hurdle.
3. Emergence of AI/ML Observability
As the AI hype cycle has accelerated, we’ve seen an explosion of new tools and capabilities to support the end-to-end model development lifecycle, from experimentation to production deployment. Critical to the continued advancement of state-of-the-art models, model monitoring tools help validate the integrity and reliability of AI/ML applications, particularly as production models and their training datasets evolve over time.
In addition to monitoring more traditional health metrics, like CPU utilization and response time, emerging model monitoring platforms assess additional concerns. This includes model performance (e.g., recurring checks against a defined evaluation metric, such as PPV), drift (e.g., detecting statistical variations between models at training time vs. production) and data quality (e.g., cardinality shifts, mismatched types).
An evolving aspect of this domain we are particularly excited about is LLM observability. This discipline builds on traditional ML monitoring to capture important signals related to building, tuning and operating LLMs in the wild. For example, companies like Weights & Biases support LLM tracing, which enables teams to visualize and monitor prompt inputs, intermediary predictions, token usage and more as transactions execute along a multi-staged LLM chain. Additionally, next-gen product analytics tools are coming to market from companies like Aquarium Learning, which provide visibility into user interactions with LLM-powered interfaces (e.g., correlating common topics and user behaviors with product metrics to prioritize backlogs and inform roadmaps). Finally, we see platforms expanding to provide pattern identification and drift detection of unstructured data entities, such as vectorized embeddings, to improve the accuracy of RAG-based workflows.
4. Industry-Wide Standardization and OpenTelemetry
Historically, the observability market has been dominated by incumbents with proprietary data formats. This has created a ‘Hotel California’ lock-in scenario, forcing organizations to integrate and administer a complex universe of disjointed monitoring solutions.
Over time, the community recognized the limitations of proprietary data formats and began collaborating on open standards, with the OpenTelemetry project leading the charge. OpenTelemetry (aka OTel) is an open standard set of wire specs, APIs, SDKs, tooling and plug-ins for collecting and routing telemetry data from modern systems in a vendor-neutral way. It began as two competing projects (OpenTracing from CNCF & OpenCensus from Google) aimed at providing a vendor-neutral distributed tracing API. The two merged in 2019 and have since added support for metrics and structured logs. OTel is now the second most active CNCF project after Kubernetes, and we’ve seen both emerging startups and incumbents align to the spec in the face of customer demand for standardization and cross-platform compatibility.
That said, while the community has generally embraced OTel, many enterprises are still early in the shift, particularly given the complexity of reinstrumenting legacy applications and converting existing runbooks tied to incumbent monitoring solutions. Nevertheless, we are excited to see OTel momentum continue in 2024 and, in particular, efforts by the CNCF to standardize query languages, improve OTLP (OpenTelemetry Protocol) data compression and introduce CI/CD telemetry support. We are also excited to watch the evolution of other open standards, including the Open Cybersecurity Schema Framework (OCSF), which attempts to set an industry-wide standard format for logging common security events.
5. Bring-Your-Own Storage Backend
We believe an architectural shift is underway amongst many next-gen o11y tools, where data warehouses are becoming “the new backend,” and providers aim to fully decouple their storage and compute layers. This separation enables each infrastructure tier to be scaled independently and according to individual capacity requirements. By eliminating proprietary (and cost-prohibitive) indexers, this approach also unlocks optionality at the storage layer. In turn, this enables customers to ‘bring-their-own’ preferred database and storage solutions, sweat existing assets and achieve more granular control over data residency and access.
Providing this sort of interoperability is easier said than done, particularly when you consider the diverse performance characteristics, schema models and indexing styles of different storage engines. Observability platforms like Coralogix are a great example. They have architected their visualization and analytics engines to run across ‘cheap and deep’ storage solutions such as AWS S3.
6. Visibility into the CI Pipeline
As software projects scale, both in terms of code base size and number of engineering resources, the complexity of their underlying CI/CD pipelines often expands. Slow build times, flaky tests and merge conflicts can delay releases, divert resources away from new feature development and increase infrastructure costs. These challenges are further amplified in the case of monorepos, where each commit has the potential to trigger build and test harnesses spanning an outsized portion of the total code base.
Engineering efficiency platforms such as Jellyfish are providing visibility into the end-to-end performance of the SDLC (e.g., cycle time, change lead time) in alignment with standard frameworks like DORA. When deviations from baseline occur, CI and test analytics solutions from companies like CircleCI (a Sapphire Ventures portfolio company) provide a next-level down of details to help pinpoint and remediate the source of a given bottleneck. They collect and alert on key metrics such as job durations and CI infrastructure utilization and can also detect flaky tests through historical analysis over multiple runs. Emerging startups like Trunk.io are collecting and then auto-commenting PRs with relevant performance data, while also providing live debug capabilities (e.g., pause jobs, SSH to ephemeral CI runners) to avoid costly restarts.
Though CI analytics and engineering intelligence tools are often considered outside the scope of ‘traditional’ monitoring or systems observability, we feel these tools are providing much-needed visibility to developers - and are indicative of a broader expansion of this category to cover all personas and phases of the product life cycle.
7. Linking Business Outcomes to Systems Data
Product experience monitoring and system monitoring tools have traditionally operated in silos. More recently, we’ve seen a push to converge these domains to understand the correlation between end-user behaviors and system-level signals. For example, product-level metrics might reveal sparse user engagement with a given feature. To enrich this metric, system monitoring tools might uncover that the feature’s underlying code suffers from high error rates or slow response times. Conversely, infrastructure monitoring tools might detect utilization spikes that, when correlated with product analytics, highlight problematic user behaviors, such as overusing a particular feature.
We see several players pushing the envelope here, augmenting Digital Experience Monitoring & Analytics with error tracking and other APM-focused metrics to provide a deeper understanding of specific user groups and their interactions with the system. Similarly, feature management platforms such as Statsig and Unleash integrate with observability platforms to better correlate feature adoption with system-level performance signals.
Exploring the Observability Market Opportunity
The growth of the o11y market, from $61B in 2019 to $105B in 2023, shown in the chart below, demonstrates the magnitude of the opportunity for transformative outcomes in this space. We’ve identified $10B+ in aggregate ARR across both public and private pure-play observability vendors (excluding solutions from both hyperscalers and other large/broad tech incumbents - but including the Splunks of the world), highlighting a thriving and enticing market for a next-generation of startups to conquer.
Expansion of the o11y market has been heavily influenced by digital transformation, which has made the data collected from applications more critical than ever. This is demonstrated by the sector’s adaptability to market shifts - navigating the initial surge of ‘growth at all costs’ IT spend during the height of COVID to the more balanced, cost-conscious spend patterns of 2022 and beyond. The market's ability to normalize and rebound post-correction underscores its resilience and strategic importance in the digital era.
Navigating the Observability Market Landscape
In the map below, we’ve attempted to sort companies based on initial and/or historical market focus. However, many leading providers have and will continue to evolve into end-to-end platforms, enabling them to compete across a majority of the functional buckets we’ve outlined. Grafana, for example, is placed in the visualization bucket (where they are the de-facto standard), but has since expanded into a full-stack observability platform, offering APM, log analytics, infra monitoring, incident response and more.
Similar platform expansions are occurring not just on individual features but also cross-domain, specifically with the convergence of IT monitoring and cybersecurity. For example, Datadog focused initially on cloud infrastructure monitoring but has since expanded into the security realm with SIEM, CSPM and CWPP. Part of the reason for this convergence, in addition to finding new avenues for revenue growth, is that security and observability tools can often take advantage of overlapping data sets and deploy similar host agents for collection. By merging these capabilities, customers can avoid agent sprawl, achieve vendor consolidation objectives and derive unique insights from the combined data sets.
*Companies are categorized by initial focus area, and many play in multiple buckets. We’ve struggled with categorization so have decided to bucket focus based on where they’ve started.
Below, we’ve provided a brief definition of the functional categories outlined in this map:
Digital Experience Monitoring & Analytics solutions observe the availability, performance and quality of end-user experiences. DEM solutions seek to model the user journey, tracking actions such as rage clicks, heat maps, individual element response times and more. Players in the space typically offer capabilities across front-end monitoring, error tracking and session replay.
Synthetic Monitoring platforms orchestrate large fleets of probes that perform scripted interactions with both web properties and API endpoints. These synthetic probes are used to provide early warning and detection of performance degradations and other functional application issues. Tools like Catchpoint and Checkly automate these simulations across multiple locations and geographies, with the capability to deploy private agents in support of remote edge sites and private data centers.
Application Performance Monitoring (APM) solutions provide a detailed view of an app’s dependencies, traces of user flows and measurement of microservice level performance through to business KPIs. APM platform leaders such as Honeycomb provide robust transaction tracing captured in the form of structured, OTel-compliant spans coupled with a high cardinality storage engine. Other APM platforms focus more on code-level monitoring, including Sentry, which, in addition to tracing, offers a polyglot SDK for capturing errors and associated stack traces, local variables and more.
ML Model Monitoring and LLM Ops tools help ensure the integrity and reliability of AI/ML applications, particularly as production models and their training datasets, evolve over time. Model monitoring platforms capture traditional health metrics (e.g., CPU utilization and response time), along with ML-specific concerns around model performance (e.g., evaluating PPV), drift (e.g., statistical variations between training and prod) and data quality. Additionally, leading MLOps solutions have begun to expand into LLM application tracing, providing visibility into the latency, cost and accuracy of ‘multi-staged’ LLM transactions.
Log Analytics platforms collect, aggregate and analyze event data from various sources, including servers, applications, networks and devices. Emerging players in this category include Coralogix, which provides native data pipeline capabilities and is capable of running atop cheap and deep storage solutions.
Infrastructure Monitoring solutions collect and analyze metrics in order to monitor and alert on the availability and resource utilization of physical and virtual assets (e.g., servers, containers, network devices, databases, hypervisors and storage). Infrastructure-focused observability tools include Chronosphere which differentiates in part with its highly performant M3 backend for Prometheus, which was developed by the founding team during their time Uber.
Network Performance Monitoring (NPM) leverages a combination of data sources (e.g., packet captures and device logs) to provide a holistic view of network performance across hybrid multi-cloud environments. These platforms provide diagnostic workflows and forensic data to identify the root causes of network performance degradations and overall reliability.
CI/CD Visibility solutions offer enhanced visibility into the end to end software development process, including job durations, CI infrastructure utilization, flaky test detection and more.
AI-Ops & Incident Response tools leverage machine learning to detect anomalies, correlate events and determine the root cause of issues. Downstream, these platforms orchestrate incident response processes with hooks into automation tools capable of troubleshooting and remediating issues. Platforms like Moveworks offer intelligent knowledge search and automated issue resolution via LLM-powered virtual agents, while startups like Rootly and Incident.io provide a framework for defining runbooks, collating relevant signals and alerts, centralizing team collaboration, tracking steps to resolution and reporting post-mortem.
Observability Pipelines are solutions that manage the collection, processing, enrichment, transformation and routing of telemetry data from source to destination. Solutions in this space are offered as either standalone proxies (e.g., Cribl) or as a subcomponent of a broader o11y suite (see Coralogix’s Streama or Chronosphere’s acquisition of Calyptia).
Visualization tools enable users to create and share dynamic graphs and telemetry dashboards, integrating data from multiple sources to provide a unified view of a system. Grafana has emerged as the de-facto standard query and visualization layer for modern applications, providing a single pane of glass that unifies data across heterogeneous platforms and environments.
Security Information and Event Management (SIEM) solutions aggregate and analyze security event data from various sources, identifying patterns that might represent a security threat or active breach, including complex attacks that may not be detected by signature-based systems. Advanced tools in this domain, such as Exabeam and Anvilogic, offer sophisticated analytics, AI-driven threat detection and downstream incident response.
Charting the Future of Observability
With the evolution of application architectures, the ever-increasing growth in telemetry volumes and the availability of powerful new AI models, the o11y market is primed for disruption. We see significant opportunities for startups to disrupt the status quo and push the more established players to adapt and keep pace. They will look to create space with more performant and cost-effective hosting models, simplified user experiences, intelligent analytics and coverage across a wider breadth of personas and workload types. And while this blog highlighted some of the more impactful trends shaping this market, we're eager to see how the landscape evolves and which new entrants and capabilities emerge in 2024 and beyond.
If you’re working on a game-changing new observability platform or have come up with a novel approach to this space, we’d love to hear from you! Feel free to drop us a line at casber@sapphireventures.com, slade@sapphireventures.com and carter@sapphireventures.com.
And a special thanks and shoutout to Amit Agarwal, Amol Kulkarni, Christine Yen, Martin Mao, Milin Desai, Wailun Chan and Yanda Erlich for reading drafts of this post.
Disclaimer: Nothing presented within this article is intended to constitute investment advice, and under no circumstances should any information provided herein be used or considered as an offer to sell or a solicitation of an offer to buy an interest in any investment fund managed by Sapphire Ventures (“Sapphire”). Information provided reflects Sapphires’ views as of a time, whereby such views are subject to change at any point and Sapphire shall not be obligated to provide notice of any change. Companies mentioned in this article are a representative sample of portfolio companies in which Sapphire has invested in which the author believes such companies fit the objective criteria stated in commentary, which do not reflect all investments made by Sapphire. A complete alphabetical list of Sapphire’s investments made by Its direct growth and sports investing strategies is available here. No assumptions should be made that investments described were or will be profitable. Due to various risks and uncertainties, actual events, results or the actual experience may differ materially from those reflected or contemplated in these statements. Nothing contained in this article may be relied upon as a guarantee or assurance as to the future success of any particular company. Past performance is not indicative of future results.
A truly comprehensive view - very insightful.
Great map, congrats!