A 4.8 star average doesn't mean what you think it does
A 4.8 on Google Reviews sounds fine. Maybe it is. But the number is measuring something different from what most people think it is.
Reviews are voluntary. Nobody asks every customer to leave one. The people who do had an experience strong enough to push them to act -- exceptional service, or a problem that went unresolved and stuck with them. The people who had a perfectly fine time, which is most of your customers, go home and don't think about it again.
The distribution you're averaging over isn't your customer base. It's a self-selected slice, weighted toward the extremes.
Leaving a review means opening an app, finding the business, writing a sentence or two, hitting submit. That friction filters out anyone who isn't particularly motivated, regardless of what they thought. The people who bother tend to be more comfortable online, more inclined to feel their opinion is worth recording. It's a personality signal more than a quality one.
It also filters by timing. Someone who had a bad experience but resolved it by phone probably won't review. Someone who had the same bad experience and heard nothing will. So even within the dissatisfied group, you're getting a skewed version.
Variance tells you more than the average. Two businesses at 4.2 stars can be in completely different situations. One might have 200 reviews clustered tightly around 4 -- steady, consistent, probably fine. The other might have 80 reviews split between people who loved it and people who really didn't, with almost nothing in between. Same number. The underlying picture looks nothing alike.
Which matters, because the fix for each is different. The business with consistent 4s probably needs incremental improvement. The one with polarised reviews has a positioning or expectation problem -- some category of customer is getting something that doesn't match what they expected. The aggregate number hides that. The distribution shows it, but only if you actually look.
A 4.8 also can't tell you what's happening right now. Reviews accumulate over time. Your average includes experiences from two years ago. Last quarter might be significantly worse, but the number won't show it until enough recent reviews pull it down -- which takes a while, and by then the damage is already done. A restaurant that changed its kitchen team six months ago, a hotel that renovated its rooms, a software product that shipped a buggy update -- none of that shows up cleanly in a trailing average. The score reflects the past.
The number you see in search results drives clicks. It's not the same number that tells you what's actually happening inside your business right now -- and most businesses treat them as though they are.
Structured feedback works differently because it asks everyone, not whoever volunteers, at a moment close to the actual experience. The responses from the quiet middle tend to be more honest and more specific than the ones from people motivated enough to open Google Maps on their own. You hear from people who had ordinary experiences -- what worked okay, what was slightly off, what they'd have mentioned to a friend if someone had asked. That's where the actual texture of your business lives. What 300 feedback responses actually look like goes into what that data looks like in practice.
Tools like Qria let you set that up -- structured questions, timed to the experience, sent to everyone rather than whoever felt like it that day.
SaaS products hit a version of the same problem with cancellation dropdowns. Most users click through without engaging, and the ones who do answer aren't representative of the ones who didn't. Users who cancel without saying why covers that. NPS runs into the same selection problem from a different direction.