Net Promoter Score (NPS) has been one of the most widely adopted customer-loyalty metrics since Fred Reichheld introduced it in 2003. Yet in 2024, academic literature has accumulated substantive methodological critiques, and practitioners continue to grapple with operational questions: "Is my score good for the industry?" "Can we compare across regions?" "Is NPS enough on its own?"
This guide pulls together the NPS basics, recent peer-reviewed critiques, industry benchmarks, and operational judgment criteria — including considerations for Japan-based operations. The Kicue editorial team has cross-referenced recent journal articles, vendor benchmarks, and public practitioner commentary to produce a grounded, decision-oriented reference.
1. The Basics — How NPS Is Calculated
NPS uses an 11-point (0–10) "recommendation intent" question, bucketed into three groups:
- Promoters: 9–10
- Passives: 7–8
- Detractors: 0–6
NPS = % Promoters − % Detractors
The resulting score ranges from -100 to +100. Note that Passives are excluded from the calculation, and the 11-point granularity is compressed into three categories — two structural choices that sit at the heart of the academic critiques covered next.
2. Structural Limits — Key Academic Critiques
A coordinated body of 2024 peer-reviewed work has outlined specific methodological concerns with NPS. Teams deploying NPS benefit from understanding these limits before committing to the metric.
Information loss from 11-to-3 collapse
Dawes (2024) in the International Journal of Market Research catalogs the structural issues with NPS calculation:
- The cutoff points dividing Promoters, Passives, and Detractors are arbitrary
- Passives (7–8) are excluded entirely from the score
- Compressing an 11-point scale into three categories discards response variance
This compression introduces extra variability in NPS compared to a straightforward mean-based approach on the same scale.
Sales growth correlation is conditional
The popular claim that "higher NPS drives higher growth" does not hold unconditionally. Research published in the Journal of the Academy of Marketing Science (2021) empirically showed that NPS:
- Has meaningful predictive value only for near-term sales growth
- Predicts better when measured across the broader prospect universe, not just existing customers
- Does not consistently outperform alternative customer-feedback metrics
In short, "high NPS → company grows" is not an academically supported causal statement on its own.
Supplement NPS with other metrics
Müller, Seiler, and Völkle (2024) show empirically that combining NPS with emotion-based metrics improves loyalty prediction. Bettencourt and Houston (2024) similarly argue for multi-metric operations rather than NPS in isolation.
The prevailing 2024 academic direction: treat NPS as one signal in a composite measurement system, not as a standalone KPI.
3. Industry Benchmarks — What Vendor Data Shows for 2025
"Is our score any good?" is an inevitable question. Several benchmark reports from customer-experience vendors are publicly available. These are industry reference points based on each vendor's client base, not academically designed studies — useful as rough orientation, not as validated ground truth.
B2B SaaS benchmarks
Multiple vendor-published 2025 benchmarks cluster B2B SaaS in the +31 to +41 range:
- Retently (2025) reports B2B Software & SaaS averaging +41
- CustomerGauge puts SaaS around +36
- NPSpack (2025) analyzed 500+ SaaS companies and found average +31
Different methodologies produce differences of 5–10 points. The fact that multiple vendors converge on the same range gives a defensible reference band: roughly +30 to +40 for B2B SaaS.
Commonly cited "good score" thresholds
Cross-referencing vendor commentary (e.g., ChurnWard, Survicate), the following interpretation is widely shared in the English-language market:
| Range | Typical interpretation |
|---|---|
| Below 0 | Needs improvement |
| 0–30 | Average |
| 30–50 | Good |
| 50–70 | Excellent |
| 70+ | Top tier |
Keep in mind this interpretation is anchored in Anglo-American markets. The next section covers why you need different expectations for Japan.
4. Operating in Japan — Dealing with Structurally Lower Scores
NPS scores measured in Japan tend to come in meaningfully lower than Anglo-American benchmarks. This observation is repeatedly noted by multiple Japan-based NPS operators.
Multiple Japanese operators report the pattern
- NTT Com Online discusses cultural response patterns behind low Japanese NPS scores
- EmotionTech notes that about one-third of Japanese responses cluster on "5," and that 9–10 ratings typically require substantial emotional investment in a product
- Qualtrics presents cross-country NPS comparison data across 18 nations/regions
This is industry operational knowledge, not peer-reviewed structural evidence. But the convergence of multiple independent operators on the same pattern is meaningful reference material when deciding whether to apply global comparisons to Japanese data.
Realistic score ranges in Japan
Synthesizing Japanese operator commentary (e.g., EmotionTech), what counts as a "leading company" NPS in Japan is often cited in the -20 to 0 range. Applying the global "30+ is good" standard directly to Japanese data risks rendering every company negative.
Operational decision criteria (editorial view)
Triangulating publicly available practitioner commentary, the following judgment rules have become common in Japanese NPS operations:
- Avoid absolute cross-region comparison — loading Japan data against a global KPI is structurally unfair
- Anchor on time-series and in-market competitive comparisons — "+3 pt QoQ" or "#2 in our category" are more honest
- Analyze Detractor free-text verbatims closely — the absolute score matters less than actionable diagnoses
- Run relational (annual) and transactional (post-interaction) NPS together — don't choose just one
5. Never Run NPS Alone — Composite Measurement Beats a Single Metric
Both the academic critique and the Japanese operational knowledge point in the same direction: NPS belongs inside a composite measurement system, not on its own KPI dashboard.
Commonly paired metrics
- CSAT (Customer Satisfaction): post-interaction satisfaction, good for short-cycle feedback
- CES (Customer Effort Score): effort to complete a task, strong for support/ops diagnostics
- Retention / Churn Rate: behavioral reality, powerful complement to attitudinal NPS
- Emotion-based metrics: shown in 2024 research to lift NPS predictive power
Treat NPS as a snapshot of sentiment, and behavioral metrics (retention, CES) as the reality check. Combine them with solid questionnaire design fundamentals for a system that actually informs decisions.
6. Building NPS Programs in the Survey Tool Kicue
Kicue ships with the features typical NPS programs require:
- NPS (0–10 scale) question type — one-click setup (question type reference)
- Segment-level cross-tabulation — use URL parameters or attribute questions as axes to compare Promoter / Passive / Detractor distributions across segments (cross-tabulation)
- Detractor follow-up by design — use display conditions to show a "why?" free-text only for respondents scoring 0–6, capturing detractor verbatims without cluttering the promoter flow
- URL parameter integration with external systems — if your CRM or email platform appends customer IDs or segment attributes to the NPS public URL, Kicue automatically binds them to the response (URL parameter docs)
- Response-rate lifting features — mobile optimization, skip logic, progress bar baked in
Upload a questionnaire file, and the platform auto-generates the question structure, branching logic, and follow-up design needed for NPS operations.
Choosing the right tool — Free plan limits, branching support, AI capabilities, and CSV export vary widely across tools. See our free survey tool comparison to find the right fit for this approach.
Recap
Operational checklist for NPS programs:
- Understand the structural limits — 11-to-3 collapse, arbitrary cutoffs
- Remember the sales-growth link is conditional — near-term horizon, broader prospect universe
- Use vendor benchmarks as reference, not ground truth — B2B SaaS sits in +30 to +40 per multiple vendors
- Account for regional patterns — Japan's "good" zone is often cited as -20 to 0 per Japanese operators
- Never run NPS alone — pair with CSAT, CES, retention, and emotion metrics
NPS is seductive because it produces a single clean number. But the number becomes genuinely useful when it's read across three axes: time series, segments, and complementary metrics.
References (13)
Academic & peer-reviewed research
- Dawes, J. G. (2024). The net promoter score: What should managers know? International Journal of Market Research.
- Bettencourt, L. A., & Houston, M. B. (2024). The Untested Assumption: Can a Net Promoter Study Be Used to Improve Net Promoter Score? International Journal of Market Research.
- Müller, S., Seiler, R., & Völkle, M. (2024). Should Net Promoter Score be supplemented with other customer feedback metrics? International Journal of Market Research.
- The use of Net Promoter Score (NPS) to predict sales growth (Journal of the Academy of Marketing Science, 2021).
Industry benchmarks & vendor commentary
- Retently: What is a Good Net Promoter Score? (2025 NPS Benchmark)
- CustomerGauge: SaaS NPS Benchmarks 2025
- NPSpack: SaaS NPS Benchmarks 2025
- Survicate: NPS Benchmarks 2025
- ChurnWard: What Is a Good NPS Score?
Japan-focused operator commentary (treated as industry knowledge)
Design NPS programs efficiently with Kicue — a free online survey tool that combines NPS, CSAT, and CES into one operational stack.
