Published — 9 min read
Data pipelines fail in ways that software systems do not. A web service either responds or it does not. A data pipeline might technically complete successfully while quietly producing wrong data for weeks. The ingestion pipeline ran. The transformation succeeded. The dashboard loaded. But the revenue figure on the executive dashboard was off by 12% because a currency conversion table was not refreshed after a rate change, and no one caught it until an analyst noticed the discrepancy in a quarterly review.
Observability for data pipelines requires a different approach than observability for software services because the failure modes are different. Correctness failures — pipelines that run without errors but produce wrong outputs — are as important as availability failures — pipelines that do not run at all. The metrics that matter for data observability track both dimensions, along with several others that are unique to data systems.
Data observability frameworks have converged on five categories of signals that together provide comprehensive visibility into pipeline health. Understanding each pillar and its corresponding metrics is the foundation for building an effective monitoring program.
Freshness: Is data being updated as expected? Freshness monitoring tracks the age of each dataset relative to its expected update schedule. A table that should be updated every hour and was last updated six hours ago is stale, even if the data it contains is correct. Freshness is often the first signal that a pipeline failure has occurred: the pipeline stopped running before any quality checks could catch it.
The most actionable freshness metric is time since last update (TSLU) compared to expected update interval. Alerting when TSLU exceeds 1.5x the expected interval provides early warning of pipeline delays before they become SLA violations. More sophisticated freshness monitoring tracks expected update time windows (this table should update between 06:00 and 07:00 UTC daily) rather than just intervals, catching pipelines that ran late rather than not at all.
Volume: Does the data volume match expectations? Volume monitoring compares the number of rows added or updated in each processing window against historical norms. A significant deviation from expected volume — either too few or too many rows — is one of the most reliable indicators of upstream data issues. A daily transaction table that normally receives between 1.8M and 2.2M rows and receives only 350,000 rows has clearly experienced an upstream problem, even if the 350,000 rows that arrived are all valid.
Volume monitoring requires establishing baselines that account for temporal patterns. Week-over-week comparisons are more meaningful than day-over-day for datasets with strong weekly seasonality. Month-over-month comparisons with seasonal adjustment are more meaningful for monthly aggregated data. Naive volume monitoring that ignores these patterns generates constant false positives.
Schema: Has the structure of the data changed unexpectedly? Schema monitoring detects column additions, removals, renames, and type changes that occur without going through a change management process. These changes are a persistent source of silent pipeline failures: a rename upstream causes a downstream join to produce NULL values instead of the expected matches, inflating null rates and deflating record counts without generating any explicit errors.
Schema monitoring is binary: either the schema matches the expected definition or it does not. The response to schema changes should also be binary: quarantine and investigate before allowing changed-schema data to flow downstream. The cost of a brief data delay is always lower than the cost of allowing schema-incompatible data to contaminate production analytics.
Distribution: Are the statistical properties of the data within expected ranges? Distribution monitoring goes beyond column-level null rates and range checks to track the full statistical profile of key metrics over time. Is the mean value of this column within 2 standard deviations of its historical mean? Is the percentage of NULL values within expected bounds? Is the cardinality of this categorical column stable?
Distribution shifts that fall outside control limits are often the first observable signal of business events that affect data — a new product launch changing the product ID cardinality, a system migration changing the encoding of a status field, a pricing change shifting the distribution of order values. These are not errors in the traditional sense, but they require investigation to determine whether the shift reflects a real business change or a data quality problem.
Lineage: What is the complete dependency graph for this dataset, and where in the graph did a problem originate? Lineage observability answers the question "this downstream dashboard is wrong; where did the error enter the pipeline?" without requiring manual investigation of every upstream dependency. Column-level lineage that tracks exactly which source columns flow into which downstream columns enables precise root cause analysis.
The highest-value observability metrics are leading indicators — signals that predict future failures rather than detecting current ones. These predictive signals allow data teams to intervene before failures affect business users, rather than reacting after the fact.
Pipeline SLA adherence rate trend: Not just whether today's pipeline met its SLA, but whether SLA adherence is improving or degrading over time. A pipeline that meets its SLA 95% of the time but has been missing SLA 20% more frequently over the last two weeks is showing a concerning trend that warrants investigation before the next miss.
Resource utilization trend: Memory, CPU, and storage utilization for compute clusters running data pipelines. Utilization trending toward capacity limits predicts future performance degradation or failures before they manifest as SLA misses. Proactive capacity planning based on utilization trends prevents a large fraction of resource-exhaustion failures.
Upstream data delivery variance: How variable is the arrival time of upstream data that your pipeline depends on? An upstream source that normally delivers data at 05:30 UTC but has been delivering between 04:45 and 06:15 UTC over the last month is increasingly unreliable. Your pipeline that kicks off at 06:00 UTC assuming data has arrived will fail with increasing frequency. Tracking upstream delivery variance enables proactive adjustments to dependency timing.
Error rate by error type: Not just total error rate, but the breakdown by error category. A rising rate of schema validation errors often precedes a full schema change. A rising rate of network timeout errors often precedes a complete network partition failure. Understanding which error types are increasing, even when total error rates are within acceptable ranges, predicts where the next significant failure is likely to originate.
The monitoring metrics are only valuable if they produce actionable alerts that reach the right people quickly. Alerting design for data observability follows the same principles as software alerting: alert on symptoms, not causes; minimize false positive rates to prevent alert fatigue; provide context that helps the recipient diagnose the issue immediately.
An effective data observability alert includes: what failed (specific table or pipeline), which pillar triggered the alert (freshness, volume, schema, distribution), what the expected and observed values were, what downstream systems are affected (from the lineage graph), and a link to the lineage view for investigation. An alert that says "ALERT: sales_daily volume anomaly" provides far less value than one that says "ALERT: sales_daily expected 2.1M rows, received 180K rows (91% below average). Last update: 4 hours ago. Downstream affected: executive_revenue_dashboard, finance_reconciliation_view. Lineage: https://..."
Alert routing matters as much as alert content. Freshness alerts for a table owned by the Commerce domain should route to the Commerce data team, not to a shared operations channel that everyone ignores. Domain-specific routing, implemented through the data product ownership model, ensures that alerts reach people with the context and authority to fix the problem quickly.
The operational maturity of a data team is measured by its ability to make and keep data SLA commitments. Borrowing the error budget concept from site reliability engineering (SRE), data reliability teams define explicit SLOs (service level objectives) for their data products: "the daily sales table will be fresh within 2 hours of business close, with 99.5% monthly uptime."
When the error budget is consumed (when SLA misses have used up the allowed failure budget for the month), the data team shifts focus from new feature development to reliability improvements. This prioritization mechanism ensures that data reliability is treated as a first-class concern rather than being perpetually deprioritized in favor of new data product development.
Measuring and reporting on data SLA adherence to business stakeholders changes the organizational conversation about data reliability from vague complaints about "bad data" to specific, quantified commitments and accountability. Business stakeholders who know they can rely on the daily sales data being ready by 07:00 AM 99.5% of months can build workflows that depend on it. Those who have no reliability commitments cannot.
Data pipeline observability is the infrastructure investment that makes every other analytics investment more reliable. Without it, data quality issues remain invisible until business users discover them, eroding trust in the analytics program. With it, data teams catch failures proactively, resolve them quickly, and build the track record of reliability that earns the trust that makes analytics valuable. The tools exist; the patterns are well understood; the investment is justified by the risk it mitigates.
See how Dataova's integrated observability layer provides continuous monitoring across all connected data sources, alerting your team to quality issues before they surface in business-critical dashboards.