Data Analysis Should Be for Everyone
Here's something the analytics industry doesn't like to acknowledge: most businesses in the world have no meaningful access to data analysis.
Not "limited access." Not "could do better." No access. Zero.
They have data — every business does. Transaction records, customer lists, website traffic, inventory counts. The data exists. But the tools, expertise, and infrastructure required to turn that data into decisions are priced and designed for a different world. A world of dedicated data teams, cloud warehouses, and six-figure software budgets.
For the bakery owner trying to figure out which products are actually profitable, the freelance consultant trying to track client retention, the 10-person startup trying to understand their funnel — the analytics industry has nothing useful to offer. At least, nothing they can afford or operate without a specialist.
This is the problem I care most about. And I think it's the biggest unsolved problem in the analytics space.
The Privilege Problem
Data analysis in 2026 is a privilege good. Like healthcare in a broken system or education behind a paywall, it's something that everyone needs and only some people can access.
The numbers tell the story. A functional analytics setup — from data ingestion to warehouse to transformation to visualization — costs $100,000 to $500,000 per year minimum when you factor in tools and personnel. Enterprise BI platforms charge per-seat fees that make them impractical for small teams. Even "affordable" tools like Metabase or Redash require someone with SQL knowledge and the time to set up and maintain them.
The result is predictable: data-driven decision making is concentrated in well-funded companies with specialized teams. Everyone else makes decisions based on spreadsheets, intuition, and hope.
This isn't a minor inefficiency. It's a structural disadvantage that compounds over time. Companies with analytics capabilities make better decisions, which leads to better outcomes, which generates more revenue, which funds more analytics capability. Companies without it fall further behind with each decision made on incomplete information.
The rich get data-richer. Everyone else flies blind.
The Decisions Being Made in the Dark
Let me make this concrete.
A SaaS startup with 15 employees is trying to figure out which features drive retention. They have the data — user activity logs, subscription status, feature usage events. But turning that data into a cohort analysis requires either hiring an analyst ($80,000+/year), paying for a BI tool and learning to use it (months of time they don't have), or asking an engineer to stop building product and write SQL queries instead.
So they don't do the analysis. They guess. They look at what features are used most (not the same as what drives retention) and prioritize based on gut feel. Maybe they guess right. Maybe they don't. They'll never know for sure because they can't measure it.
A small e-commerce business wants to understand their customer lifetime value by acquisition channel. The data exists — order history, marketing attribution, customer profiles. But connecting those data points requires joins across multiple tables, cohort definitions, and statistical rigor that goes beyond what a spreadsheet can handle.
So they keep spending equally across all channels. They don't know that one channel produces customers who spend 3x more over their lifetime. They're leaving money on the table — potentially a lot of it — because the analysis is beyond their reach.
These aren't hypothetical scenarios. They're the daily reality for most businesses. Important decisions made with incomplete information because the tools to analyze the available data are too expensive, too complex, or both.
Why "Self-Serve Analytics" Failed
The industry has been promising "self-serve analytics" for over a decade. The pitch: give business users tools to analyze data themselves, without depending on data teams.
It hasn't worked. And the reason is instructive.
Self-serve analytics tools made the same mistake as most analytics tools: they assumed the hard part was the interface. Build a good enough drag-and-drop builder, a good enough natural language interface, a good enough no-code query tool, and non-technical users will be able to analyze data themselves.
But the interface was never the hard part. The hard parts are:
Understanding data models. Even the best interface requires users to know which tables to query, how they're related, and what the columns mean. A business user looking at a schema with tables named dim_customers, fct_orders, and stg_events doesn't know where to start. And they shouldn't have to.
Getting consistent metrics. Self-serve tools let users build their own analyses, which means they build their own metric definitions. Marketing calculates revenue one way. Finance calculates it another. Product calculates it a third way. Self-serve didn't create consistency — it created chaos with a friendly interface.
Trusting the results. When a business user builds a dashboard in a self-serve tool and the numbers look wrong, they don't debug the query. They lose trust in the tool. They go back to asking the data team (if one exists) or go back to the spreadsheet.
Self-serve analytics failed because it democratized the interface without democratizing the foundation. It gave everyone access to a query builder without giving them consistent metrics, understandable data models, or confidence that their results were correct.
The solution isn't a better interface. It's a better foundation.
What Actually Accessible Analytics Looks Like
If I had to describe what truly accessible analytics looks like, it would have these properties:
Metrics are pre-defined and trustworthy. Users don't need to know SQL or understand the data model. They need to know what they want to measure — revenue, retention, conversion, whatever — and the system needs to already know how to calculate it correctly. The metric definitions are the foundation. Everything else is presentation.
The data stays where it is. Asking a small team to set up a data warehouse before they can analyze anything is like asking someone to build a highway before they can drive to the store. The tool should work with whatever database they have — PostgreSQL, MySQL, even a SQLite file or a spreadsheet.
Time-to-value is measured in minutes. Not days. Not weeks. Not "after the implementation consultant finishes the initial setup." Connect your database, define a few metrics, start getting insights. If the tool can't deliver value in a single sitting, it's not accessible — it's just less expensive.
AI bridges the expertise gap. This is where AI actually helps, as opposed to the Text2SQL parlor trick. When metrics are well-defined in a semantic layer, AI agents can work with them effectively. A user can ask "why did revenue drop last month?" and the AI can explore the defined metrics — checking by segment, by region, by product — and surface a meaningful answer. Not because the AI is brilliant, but because the foundation gives it the right building blocks to work with.
It costs what small teams can actually pay. This might be the simplest requirement and the hardest for the industry to accept. Accessible doesn't mean "enterprise pricing with a startup discount." It means pricing that a 5-person company can budget for without a committee meeting.
The Role AI Actually Plays
There's a lot of hype about AI democratizing analytics. Most of it misses the point.
AI doesn't democratize analytics by letting anyone write SQL in English. That's a slightly more accessible version of the same broken paradigm. If the underlying data is messy, the metrics are undefined, and the business logic is scattered across a dozen systems, asking questions in English instead of SQL doesn't help. You'll just get wrong answers in a friendlier format.
AI democratizes analytics by reducing the expertise required at every layer:
At the metric definition layer, AI can help non-technical users define metrics by suggesting calculations based on their data schema. "It looks like you have a transactions table with an amount column — would you like to define a revenue metric?" is more helpful than a blank form that expects users to know what a measure is.
At the exploration layer, AI agents working with well-defined metrics can do the iterative analysis that previously required a skilled analyst. Not perfectly — not yet — but well enough to surface basic insights that would otherwise go undiscovered.
At the interpretation layer, AI can explain results in plain language. Not just "revenue was $1.2M" but "revenue was $1.2M, which is 8% below the previous quarter, primarily driven by a drop in the enterprise segment."
None of this works without a proper foundation. AI is an amplifier, not a replacement. It amplifies whatever foundation you give it — good or bad. Give it well-defined metrics and a clean semantic layer, and it amplifies clarity. Give it raw tables and undefined business logic, and it amplifies confusion.
Why This Matters Beyond Business
I've been talking about businesses, but the implications go further.
Non-profits that can't afford analytics make less efficient use of donations. Government agencies without data capabilities make policy decisions on outdated or incomplete information. Schools that can't analyze student performance data miss interventions that could change outcomes.
Data literacy advocates talk about teaching people to read data. That's important. But it's not enough to teach people to read if you don't give them access to books. The analytics tools are the books. And right now, they're locked behind a paywall that excludes most of the people who would benefit from them.
This isn't just a market opportunity. It's a genuine inequity. The organizations with the least resources — the ones that most need to make every dollar count, every decision count — are the ones least able to afford the tools that would help them do it.
The Path to Getting There
I'm not going to pretend this is easy. Making analytics genuinely accessible is a hard problem. If it were easy, someone would have solved it already.
The hard parts are:
Making it simple without making it simplistic. Real business data is messy. Real business questions are nuanced. A tool that only handles simple cases isn't accessible — it's limited. The challenge is handling real-world complexity through the right abstractions, not by exposing the complexity to the user.
Making it cheap without making it unsustainable. Tools need revenue to survive. Free tools die or get acquired and ruined. The challenge is finding a pricing model that's accessible to small teams while sustainable for the company building the tool.
Building trust from the bottom up. Enterprise tools get trust through sales processes, certifications, and brand recognition. An accessible tool used by a 5-person startup doesn't have any of that. Trust has to come from the product itself — transparent code (open source helps here), correct results, and honest communication about limitations.
I don't have all the answers. We're still figuring out the right balance. But I know the direction is right because I've been on the other side — the side where you need data analysis, you know the data exists, and the path between the two is too long, too expensive, and too complex.
Data analysis shouldn't be a privilege. The data already exists. The questions already exist. The only thing missing is a bridge between the two that doesn't require a six-figure budget to cross.
That bridge should exist for everyone. Not just the companies that can afford it.