Jim Barker
jim-barker
Thought leadership
-
March 17, 2026

The House of Data Series: Data Consumption

25 min read

This paper focuses on consumption patterns, consumer behavior, and the feedback loops between data teams and business users. It does not cover pipeline architecture, quality monitoring tooling, or data governance frameworks in depth — those are addressed in the DataOps, Data Quality, and Compliance whitepapers.

Jim Barker
Get Data Insights Delivered
Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.
Join The AI Trust Summit on April 16
A one-day virtual summit on the controls enterprise leaders need to scale AI where it counts.
Get the Best of Data Leadership
Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Get the Best of Data Leadership

Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Stay Informed

Sign up for the Data Leaders Digest and get the latest trends, insights, and strategies in data management delivered straight to your inbox.

Get Data Insights Delivered

Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.

House of Data Series

Every strong data program is built like a house. Data Architecture forms the foundation — the platforms, pipelines, and operating model that everything else depends on. Seven domain pillars rise from that foundation, each one essential to a complete data program: Data Quality, Privacy, Data Security, DataOps, Compliance, Data Enablement, and Data Consumption. Data Literacy runs across all seven as a connecting beam, ensuring people at every level can read, interpret, and act on data. At the top, People & Leadership sets the direction, accountability, and culture that holds the whole structure together.

This series of whitepapers covers each component of the House of Data in depth. Each paper was written by a practitioner with direct experience in that domain. Together, they form a practical guide to building data programs that earn — and keep — trust.

Data Leadership Data Literacy Data Quality Privacy Data Security DataOps Compliance Data Enablement Data Consumption Data Architecture

This paper covers Data Consumption — the seventh pillar of the House of Data, and the reason all the other work gets done. Every pipeline built, every quality check configured, every governance policy written exists in service of this: getting trusted data into the hands of people and systems that use it to make decisions. This paper covers the five forms data consumption takes, how consumers can engage more effectively with the capabilities available to them, and how to bring research-level critical thinking to the data you rely on every day.

Data Consumption

There is a large amount of work that goes into making data available and trusted. Pipelines are built, schemas are governed, quality checks are configured, access controls are enforced, and documentation is written. All of it exists for one reason: consumption. Someone needs to use the data to do something — make a decision, automate a process, respond to a condition, or answer a question.

That sounds obvious, but in practice many data programs lose sight of it. Teams optimize for pipeline reliability, for coverage of monitoring rules, for completeness of the data catalog. Those are all valuable. They're also means, not ends. The measure of a data program is whether the people and systems consuming its outputs are making better decisions than they would without it.

This paper is a reminder of what the outcome of all that work looks like — the capabilities that should be available to consumers, the ways consumers can engage with those capabilities more effectively, and the critical thinking skills that separate data-literate users from people who treat whatever the dashboard shows as automatically correct.

Capabilities 

Data programs produce outputs in five distinct forms. Building capabilities across all five is strongly recommended — each serves a different audience, a different decision-making speed, and a different level of analytical depth.

Data programs produce outputs in five distinct forms. The table below describes each consumption type, its trade-offs, and the primary business value it delivers. Building capabilities across all five is recommended — each serves a different audience and decision-making speed.

Data programs produce outputs in five distinct forms. The table below describes each consumption type, its trade-offs, and the primary business value it delivers. Building capabilities across all five is recommended — each serves a different audience and decision-making speed.

CAPABILITY DESCRIPTION PROS CONS #1 BENEFIT
AI Capabilities that use AI to generate answers, surface patterns in data, and show users what isn't visible through manual review alone. Speed and creativity — hidden patterns surface faster than any manual process allows. Risk of sharing data outside the organization that needs to stay confidential. Getting better answers to the business faster so creative action can happen sooner.
Alerts Notifications triggered when something occurs that requires a business decision or action. Immediate notification that prompts a response before a problem compounds. Alert overload — if volume is too high, recipients tune them out entirely. Rapid response to conditions that matter.
Dashboards A grouped set of summarized metrics that tells a story tailored to a specific audience or persona. Illustrates organizational position at a glance, making it easy to identify where action is needed. High-level by design — typically requires drill-down to understand the cause behind a trend. Executive visibility across the metrics that matter most.
Digital Transformation Redesigning business processes to reduce manual work and improve the speed and quality of delivery for both customers and employees. Greater efficiency, improved ease of use, and faster time to delivery. Customers or employees can feel more disconnected from the process when human touchpoints are removed. Efficiency at scale.
Reports Granular exposure of a situation in summary or detailed grid format, covering whatever range of data the user needs to understand what's happening. Flexible by design — can be scoped to fit the specific need of any user or situation. Report proliferation — when there are too many, users struggle to find what they need. Tells the full story and provides the detail needed to take informed action.

Consumers should understand and leverage these capabilities, but it isn’t enough to simply use them. How they are used for understanding and actively engaging with your business is more important. The following is a chart that illustrates how increases in software capabilities along the X axis aligned with growth of the intellect of your staff will grow out stronger and stronger uses of your data. 

Think of this, as your staff gets smarter, increases their data literacy, and sees how topics such as AI provide benefit match up with the improvement in technology what kind of things can you accomplish:

Assistance

Consumption will be providing new capabilities as those are rolled out; they should be provided a set of job aids, training materials, and recordings to help the user be successful. The consumers of these capabilities need to take advantage of them. Time is invested in the building of these to help people be successful, not simply to check the box that has created something to make the project management office happy.

As teams are successful with new capabilities, data governance teams should survey the consumers to help understand what is working, what is not, and what to do differently next time.

Ask #2: As a consumer - Complete any survey sent to you by the data governance team.

As you begin to use these capabilities, leverage the assets, be open to share your experiences, but don’t be afraid to ask for help.

Ask #3: When you are stuck, ask for help!!!

Ideation: closing the feedback loop

Data leaders don't know everything the business needs. That isn't a failure of the data program — it's a structural reality. The people closest to the data, using it to do their jobs every day, have a clearer view of what's missing, what's confusing, and what would actually make a difference than any central team can develop from the outside.

Ideation is the set of channels through which consumers can surface that knowledge. Five tools are described here, each suited to a different kind of input and a different kind of user.

Data teams don't have full visibility into what consumers actually need. These five ideation channels give consumers a structured way to surface ideas, flag problems, and shape the direction of new capabilities — making the feedback loop between users and data governance teams explicit.

Data teams don't have full visibility into what consumers actually need. These five ideation channels give consumers a structured way to surface ideas, flag problems, and shape the direction of new capabilities — making the feedback loop between users and data governance teams explicit.

IDEATION TOOL DESCRIPTION
Data Governance Inbox A hosted mailbox managed by an on-call data steward that acts as a suggestion box for new ideas, concerns, and challenges from across the organization.
AI Solutions An AI assistant that accepts questions and inputs, or points users toward relevant resources — lowering the barrier to engagement for consumers who aren't sure where to start.
Survey Requests Surveys distributed to the community of practice to collect anonymous feedback on existing solutions and identify where additional assistance is needed.
Community of Practice Sessions Quarterly brown-bag sessions with open Q&A. Feedback collected in these sessions directly shapes the direction of the data program.
Facilitated "Future" Sessions Small-group focus sessions with data advocates, designed to generate forward-looking ideas. Deliberately kept small to enable honest, generative conversation.

Great ideas come from people who use the data. The channels above exist to capture those ideas systematically — but they only work if consumers actually use them. Providing ideas for new capabilities isn't an optional contribution. It's part of what makes a data program better over time.

Coaching: sharing what you know

Every data program has people who've figured things out — who know which reports to trust, how to get help when they're stuck, which data steward to call about which dataset. Those people are data coaches, whether or not they have a formal title.

Effective data programs make coaching explicit. They identify who the coaches are, make that visible to others, and create space for knowledge sharing rather than leaving it to happen by accident. When someone gets stuck, they should know who to ask. And when someone figures out something useful, the expectation should be that they share it rather than keeping it as a personal advantage.

Sharing how you use data capabilities helps your entire organization. Participating in the community of practice, writing up what works, being available when a colleague is stuck — this is how institutional knowledge compounds rather than walks out the door when individuals change roles.

Critical thinking: four research concepts for everyday data use

One of the persistent problems in data consumption is autopilot. A report is generated, a system provides it, and it's treated as correct because the system produced it. This problem is more pronounced with AI, where the confidence of the output can mask the uncertainty of the underlying data. People trust the number because a machine provided it, and machines seem authoritative.

There's a useful corrective from the field of scholarly research. Four concepts — validity, reliability, dependability, and bias — are standard tools in academic research methodology. They translate directly to how practitioners should approach any data output they're about to act on.

Not all reported data is equally trustworthy. These four concepts, drawn from research methodology, give consumers a framework for critically evaluating any report, dashboard, or AI output before acting on it.

Not all reported data is equally trustworthy. These four concepts, drawn from research methodology, give consumers a framework for critically evaluating any report, dashboard, or AI output before acting on it.

CONSIDERATION DESCRIPTION
Validity The closeness between what is actually being measured and what was intended to be measured. A retailer testing return rates with and without a promotional campaign, unaware that a vendor campaign was running simultaneously, is measuring something different than they think.
Reliability How consistently an analysis produces the same or similar results across different circumstances. A measurement of grocery store inventory rates that holds across pre-, during-, and post-COVID conditions is more reliable than one that only holds in a single context.
Dependability The steps taken to ensure information is available and correct when it's needed. Dependability is an active discipline, not a passive assumption — it requires documented processes and ownership.
Bias (reduced) Conducting analysis in a way that prevents confounding, misrepresentation, misclassification, and conclusions based on personal preference. In practice, this means not re-running an analysis until it produces a preferred result.

These aren't academic abstractions. Every analyst who has re-run a query until it produced an expected answer has encountered the bias problem. Every team that has reported on a metric without understanding what it was actually capturing has encountered the validity problem. Bringing these concepts explicitly into how consumers approach data outputs raises the standard for what counts as a sound basis for a business decision.

Voice of the customer and continuous improvement

Consumption isn't a one-way channel where data teams push outputs and business users accept them. The most effective data programs treat consumption as a continuous feedback loop — and Voice of the Customer (VOC) is one of the most practical tools for running that loop well.

VOC is a structured process from Six Sigma for collecting unfiltered input from the people using your product or service. In data program terms, that means collecting feedback from the people using reports, dashboards, and AI outputs — without leading the witness. VOC exercises that start with a biased question produce biased results. The goal is to hear what users actually experience, not to confirm what the team thinks is working.

Done right, VOC increases user satisfaction, improves analytical outcomes, and demonstrates that the data program is responsive to the people it serves. It can be as simple as a structured conversation or as involved as an observed "chairside" session — watching a user interact with an analytical product and recording what they do and where they struggle, without intervening.

Six Sigma: 5 s’s

Data professionals who handle large volumes of ad hoc requests, manage their own analytical environments, or run recurring reporting processes often struggle with organization and efficiency. The 5 S's — a lean manufacturing framework adapted for knowledge work — offer a practical structure for improving output without increasing effort.

Sort means eliminating what isn't needed: outdated reports, unused data sources, redundant queries. Straighten means organizing what remains so the right thing is easy to find. Shine means regularly reviewing and cleaning up the work environment — stale queries, broken connections, undocumented files. Standardize means building systems and procedures so the first three steps happen consistently rather than in a crisis. Sustain means maintaining that standard through continuous improvement rather than letting entropy win over time.

Teams that apply this framework to their data work typically find that a significant portion of their time is being spent on work that could be eliminated, consolidated, or made faster with minimal effort. The payoff isn't just efficiency — it's cleaner outputs and less time explaining why something is inconsistent with something else.

AI trust and data consumption

The most significant shift in data consumption in recent years is the arrival of AI as a primary consumer and a primary output channel. AI systems both consume data (training sets, live inference feeds) and produce data outputs (predictions, generated content, recommendations). Both directions introduce trust considerations that weren't present in traditional reporting environments.

On the consumption side, AI capabilities complete business processes faster and surface insights that manual analysis would miss. But they don't eliminate the need for human judgment — they change its form. Autonomous processes almost always have a level of review and approval built in, and those review steps only work if the humans involved are meaningfully engaged, not just rubber-stamping outputs.

Achieving AI Trust in consumption requires building control loops into AI-powered processes: mechanisms that allow people to raise alerts, flag concerns, or take corrective action when something looks wrong. "Human-involved" and "human-participating" aren't the same as "human-approved after a 10-second glance." The standard is active engagement from people who understand what the process is doing and are positioned to catch failures when they occur.

Bigeye’s role in consumption

Bigeye supports data consumption in three ways — two of which are invisible to the end user by design, and one that surfaces directly inside the tools they already use.

Bigeye supports data consumption in three ways — two of which are invisible to the end user by design, and one that surfaces directly inside the tools they already use.

CAPABILITY WHAT IT DOES
Preparation Identifies incidents before they reach the consumer — data that didn't load as expected, incorrect volumes, or quality issues that would otherwise surface as wrong numbers in a report. The goal is resolution before the user is ever aware a problem existed.
Provenance Provides lineage graphs for users who need to understand where data comes from and how it was transformed. For users who don't need direct access, data stewards can surface the same information on request.
Data Quality Visibility A browser extension that shows data health indicators directly within Looker, PowerBI, and Tableau — without leaving the reporting tool. Green indicates healthy data, red flags active issues, with detail available on demand.

Bigeye browser extension showing data quality health indicators within Looker, PowerBI, and Tableau.

Looker

PowerBI

Tableau

Summary

Consumption is what data programs are built for. All the architecture, quality monitoring, governance, and security work exists so that the people and systems who need data can get it, trust it, and use it to make better decisions.

That outcome doesn't happen automatically. It requires consumers who actively engage with the capabilities available to them, who provide feedback so teams know what's working, who share their knowledge with others, and who bring critical thinking to what they're looking at rather than treating the output of any system as automatically correct.

It also requires data programs that close the loop — building feedback channels, training users, making data quality visible at the point of consumption, and continuously improving based on what they hear. The five consumption capabilities, the ideation tools, the research concepts, and the quality visibility provided by tooling like Bigeye all point toward the same goal: data that gets used, and used well.

Explore the Series

Every great data program is built from the ground up.

The House of Data breaks down the ten pillars of a mature, trustworthy data organization. Click any section to explore that paper.

Data Leadership Data Literacy Data Quality Privacy Data Security DataOps Compliance Data Enablement Data Consumption Data Architecture

References

Roberts, Paula, and Helena Priest. "Reliability and validity in research." Nursing Standard, vol. 20, no. 44, 12 July 2006, pp. 41+. Gale OneFile: Health and Medicine. Accessed 16 Nov. 2025.

Avizienis, A., Laprie, J. C., & Randell, B. (2001). Fundamental concepts of dependability. Technical Report Series-University of Newcastle upon Tyne Computing Science.

Dubrova, E. (2013). Fundamentals of dependability. In Fault-tolerant design (pp. 5-20). New York, NY: Springer New York.

Voice of the Customer in Six Sigma (2025).

An Introduction to Lean 5 S’.

Chakravorty, S. S., & Shah, A. D. (2012). Lean Six Sigma (LSS): an implementation experience. European Journal of Industrial Engineering, 6(1), 118-137.

Wedgwood, I. D. (2006). Lean sigma: A practitioner's guide. Prentice Hall PTR.

share with a colleague
Resource
Monthly cost ($)
Number of resources
Time (months)
Total cost ($)
Software/Data engineer
$15,000
3
12
$540,000
Data analyst
$12,000
2
6
$144,000
Business analyst
$10,000
1
3
$30,000
Data/product manager
$20,000
2
6
$240,000
Total cost
$954,000
Role
Goals
Common needs
Data engineers
Overall data flow. Data is fresh and operating at full volume. Jobs are always running, so data outages don't impact downstream systems.
Freshness + volume
Monitoring
Schema change detection
Lineage monitoring
Data scientists
Specific datasets in great detail. Looking for outliers, duplication, and other—sometimes subtle—issues that could affect their analysis or machine learning models.
Freshness monitoringCompleteness monitoringDuplicate detectionOutlier detectionDistribution shift detectionDimensional slicing and dicing
Analytics engineers
Rapidly testing the changes they’re making within the data model. Move fast and not break things—without spending hours writing tons of pipeline tests.
Lineage monitoringETL blue/green testing
Business intelligence analysts
The business impact of data. Understand where they should spend their time digging in, and when they have a red herring caused by a data pipeline problem.
Integration with analytics toolsAnomaly detectionCustom business metricsDimensional slicing and dicing
Other stakeholders
Data reliability. Customers and stakeholders don’t want data issues to bog them down, delay deadlines, or provide inaccurate information.
Integration with analytics toolsReporting and insights

What are the five types of data consumption, and how do I know which one to build?

The five are AI, alerts, dashboards, digital transformation, and reports. They're not mutually exclusive — a mature data program should have capabilities across all five. The right starting point depends on your audience and the decision speed you're optimizing for. Executives need dashboards. Operations teams need alerts. Analysts need reports. AI is most valuable once the underlying data is trusted enough to train on.

What's the difference between a data coach and a data steward?

A data steward owns the quality and governance of a specific dataset — they're accountable for what the data contains, what it means, and whether it meets quality standards. A data coach is more about capability transfer: they're the person in your team or org who knows how to use data tools effectively, how to get help, and how to make sense of outputs. Stewards own the data. Coaches help people use it.

How should data governance teams handle feedback from consumers?

Through consistent, structured channels rather than ad hoc requests. The ideation tools in this paper — a governance inbox, surveys, community of practice sessions, facilitated focus groups — exist to make feedback collection systematic. One-off feedback gets lost. Feedback collected through repeatable channels gets prioritized, tracked, and acted on. Consumers who see their input reflected in new capabilities are also more likely to keep providing it.

about the author

Jim Barker

Director of Professional Services

Jim Barker is a lifelong data practitioner, industry thought leader, and passionate advocate for treating data as a strategic asset. With more than four decades of experience spanning data quality, governance, warehousing, migration, and architecture, Jim brings a rare blend of hands-on expertise and executive perspective to the evolving data landscape.

Jim’s journey in data began at just 14 years old. Since then, he has held leadership roles across organizations including Honeywell, Informatica, Thomson Reuters, Winshuttle (Precisely), Alation, nCloud Integrators, and Wavicle, contributing to advancements in data governance, migration methodologies, and enterprise data strategies. His work has included building global data quality programs, developing scalable governance frameworks, and driving innovation recognized across the industry.

His research and writing focus on lean data management, governance strategies, and the intersection of AI, data quality, and enterprise value creation.

Now at Bigeye as Director of Professional Services, Jim is energized by the company’s vision for data observability and its role in shaping the future of trusted data. He continues to share his perspectives through writing and speaking, aiming to elevate the conversation around data, cut through industry noise, and help organizations do data the right way.

Outside of work, Jim enjoys coaching and spending time with his family, often on the basketball court or soccer field, where many of the same lessons about teamwork, discipline, and leadership apply.

As Jim puts it: “Data matters.”

about the author

about the author

Jim Barker is a lifelong data practitioner, industry thought leader, and passionate advocate for treating data as a strategic asset. With more than four decades of experience spanning data quality, governance, warehousing, migration, and architecture, Jim brings a rare blend of hands-on expertise and executive perspective to the evolving data landscape.

Jim’s journey in data began at just 14 years old. Since then, he has held leadership roles across organizations including Honeywell, Informatica, Thomson Reuters, Winshuttle (Precisely), Alation, nCloud Integrators, and Wavicle, contributing to advancements in data governance, migration methodologies, and enterprise data strategies. His work has included building global data quality programs, developing scalable governance frameworks, and driving innovation recognized across the industry.

His research and writing focus on lean data management, governance strategies, and the intersection of AI, data quality, and enterprise value creation.

Now at Bigeye as Director of Professional Services, Jim is energized by the company’s vision for data observability and its role in shaping the future of trusted data. He continues to share his perspectives through writing and speaking, aiming to elevate the conversation around data, cut through industry noise, and help organizations do data the right way.

Outside of work, Jim enjoys coaching and spending time with his family, often on the basketball court or soccer field, where many of the same lessons about teamwork, discipline, and leadership apply.

As Jim puts it: “Data matters.”

Get the Best of Data Leadership

Subscribe to the Data Leaders Digest for exclusive content on data reliability, observability, and leadership from top industry experts.

Want the practical playbook?

Join us on April 16 for The AI Trust Summit, a one-day virtual summit focused on the production blockers that keep enterprise AI from scaling: reliability, permissions, auditability, data readiness, and governance.

Get Data Insights Delivered

Join hundreds of data professionals who subscribe to the Data Leaders Digest for actionable insights and expert advice.

Join the Bigeye Newsletter

1x per month. Get the latest in data observability right in your inbox.