“Net Promoter Score (NPS) Estimate”
Autotask PSA Datto RMM Datto Backup Microsoft 365 SmileBack HubSpot IT Glue All reports
AI-GENERATED REPORT
You searched for:

Net Promoter Score (NPS) Estimate

AI-generated loyalty analysis from SmileBack NPS surveys and CSAT ticket data

Built from: SmileBack CSAT
How this report was made
1
Autotask PSA
Multiple data sources combined
2
Proxuma Power BI
Pre-built MSP semantic model, 50+ measures
3
AI via MCP
Claude or ChatGPT writes DAX queries, executes them, formats output
4
This Report
KPIs, breakdowns, trends, recommendations
Ready in < 15 min

Net Promoter Score (NPS) Estimate

AI-generated loyalty analysis from SmileBack NPS surveys and CSAT ticket data

The data covers the full scope of Autotask PSA records relevant to this analysis, broken down by the key dimensions your team needs for day-to-day decisions and client reporting.

Who should use this: Service managers, account managers, and MSP leadership tracking customer experience

How often: Weekly for trend monitoring, monthly for team reviews, quarterly for QBRs

Time saved
Aggregating satisfaction data from survey tools and mapping it to clients takes hours. This report automates it.
Early warning
Declining satisfaction scores predict churn. Catching the trend early gives you time to act.
QBR material
Client-ready satisfaction data with trends and benchmarks for quarterly reviews.
Report categoryCSAT & Customer Satisfaction
Data sourceAutotask PSA · Datto RMM · Datto Backup · Microsoft 365 · SmileBack · HubSpot · IT Glue
RefreshReal-time via Power BI
Generation timeUnder 15 minutes
AI requiredClaude, ChatGPT or Copilot
AudienceService managers, account managers
Where to find this in Proxuma
Power BI › CSAT › Net Promoter Score (NPS) Estimate
What you can measure in this report
Executive Summary
NPS Category Breakdown
Score Distribution
NPS by Campaign
CSAT vs. NPS Correlation
NPS Score Categories Explained
Analysis
What Should You Do With This Data?
Frequently Asked Questions
Estimated NPS
Avg. Score
CSAT Positive Rate
AI-Generated Power BI Report
Net Promoter Score (NPS) Estimate

AI-generated loyalty analysis from SmileBack NPS surveys and CSAT ticket data

Demo Report: This report uses synthetic data to demonstrate AI-generated insights from Proxuma Power BI. The structure, DAX queries, and analysis reflect real MSP data patterns.
1.0 Executive Summary

Key NPS and satisfaction metrics at a glance.

Estimated NPS
87.7%
NPS proxy from SmileBack
Avg. Score
10,178
Ratings
CSAT Positive Rate
92.2%
9,385 positive out of 10,178 ticket reviews
Total Responses
272
NPS survey responses collected across all campaigns
View DAX Query - Overall NPS Calculation
EVALUATE ROW("CSATAvg", [CSAT - Average Rating], "Ratings", [CSAT - Total Ratings])
2.0 NPS Category Breakdown

How responses split across Promoters, Passives, and Detractors.

Promoters (9-10): 46 (16.9%)
Passives (7-8): 123 (45.2%)
Detractors (0-6): 103 (37.9%)
17%
45%
38%

NPS = % Promoters minus % Detractors = 16.9% − 37.9% = -21.0

View DAX Query - NPS Category Split
EVALUATE
VAR AllResponses = COUNTROWS('BI_SmileBack_Nps_Responses')
VAR Promoters = CALCULATE(COUNTROWS('BI_SmileBack_Nps_Responses'), 'BI_SmileBack_Nps_Responses'[score] >= 9)
VAR Detractors = CALCULATE(COUNTROWS('BI_SmileBack_Nps_Responses'), 'BI_SmileBack_Nps_Responses'[score] <= 6)
VAR Passives = AllResponses - Promoters - Detractors
VAR NPS = DIVIDE(Promoters - Detractors, AllResponses) * 100
VAR AvgScore = AVERAGE('BI_SmileBack_Nps_Responses'[score])
RETURN ROW(
  "Total", AllResponses,
  "Promoters", Promoters,
  "Passives", Passives,
  "Detractors", Detractors,
  "PromoterPct", DIVIDE(Promoters, AllResponses) * 100,
  "PassivePct", DIVIDE(Passives, AllResponses) * 100,
  "DetractorPct", DIVIDE(Detractors, AllResponses) * 100,
  "NPS", NPS,
  "AvgScore", AvgScore
)
3.0 Score Distribution

Response count for each score value (0 through 10).

0
17
1
4
2
16
3
13
4
9
5
21
6
23
7
57
8
66
9
22
10
24
View DAX Query - Score Distribution Query
EVALUATE
ADDCOLUMNS(
  VALUES('BI_SmileBack_Nps_Responses'[score]),
  "Count", CALCULATE(COUNTROWS('BI_SmileBack_Nps_Responses'))
)
ORDER BY 'BI_SmileBack_Nps_Responses'[score]
4.0 NPS by Campaign

Performance comparison across survey campaigns.

CampaignResponses Avg. Score Promoters Detractors NPS
Q4 Annual Survey 99 6.72 19.2% 27.3% -8.1
Mid-Year Check-in 83 6.95 20.5% 25.3% -4.8
Service Review v3 52 5.35 13.5% 55.8% -42.3
Service Review 21 4.71 4.8% 81.0% -76.2
Welcome/Exit 16 5.63 6.3% 56.3% -50.0
5.0 CSAT vs. NPS Correlation

Comparing ticket-level satisfaction with long-term loyalty.

SmileBack collects two types of feedback: CSAT ratings on individual tickets (thumbs up/neutral/down) and NPS surveys sent periodically via campaigns. Comparing the two reveals whether day-to-day satisfaction translates into long-term loyalty.

CSAT Positive Rate
92.2%
9,385 positive / 10,178 total reviews
Estimated NPS
-21
46 promoters vs. 103 detractors

The gap is significant: 92% of ticket interactions end positively, yet the NPS score sits at -21. This tells us that clients are satisfied with individual ticket resolutions but hold broader concerns about the overall relationship. The 45% passive segment (scoring 7-8) represents the biggest opportunity: they are not unhappy, but they are not enthusiastic either.

6.0 NPS Score Categories Explained

What each score range means for client behavior and retention.

Score RangeCategoryCount% of TotalTypical Behavior
9 - 10Promoter 4616.9% Actively recommends your MSP. Likely to renew and expand.
7 - 8Passive 12345.2% Satisfied but not loyal. Vulnerable to competitor offers.
0 - 6Detractor 10337.9% At risk of churn. May share negative experiences publicly.
7.0 Analysis

The data paints a clear picture: operational satisfaction is high, but strategic loyalty is low. With a CSAT positive rate of 92.2%, day-to-day ticket work meets expectations. The NPS of -21, on the other hand, signals that clients do not feel strongly enough to recommend the service.

The core issue sits in the passive segment. At 45.2% of all responses, this group scored 7 or 8. They are not dissatisfied. They are simply indifferent. Converting even a third of passives into promoters would shift the NPS from -21 to roughly +3, crossing into positive territory.

Campaign data adds context. The most recent surveys ("Q4 Annual Survey" at -8.1 and "Mid-Year Check-in" at -4.8) show improvement over older campaigns like "Service Review" at -76.2. This suggests that service quality has been trending upward, but legacy detractor responses still weigh down the overall score.

The recommended path forward: focus on the 123 passive respondents. Proactive account reviews, quarterly business reviews, and personalized follow-ups on ticket resolutions can move these clients from "fine" to "great."

8.0 What Should You Do With This Data?

Prioritized recommendations based on the NPS data.

01

Launch a Passive-to-Promoter conversion program

123 clients scored 7-8. Schedule quarterly business reviews with each, focusing on their specific pain points. A 30% conversion rate would bring NPS above zero.

02

Investigate high-detractor campaigns

The "Service Review" (NPS -76.2) and "Welcome/Exit" (NPS -50.0) campaigns carry the worst scores. Analyze what changed between these and the newer, better-performing campaigns.

03

Add a follow-up loop to NPS surveys

Only the score is captured. Adding a required comment field for scores below 7 would provide qualitative context for each detractor response.

04

Continue improving CSAT on individual tickets

The 92.2% positive CSAT rate is strong. Maintaining and publicizing this during account reviews reinforces the operational quality that NPS alone does not capture.

9.0 Frequently Asked Questions
How is NPS calculated?

NPS equals the percentage of Promoters (score 9-10) minus the percentage of Detractors (score 0-6). Scores of 7-8 are classified as Passives and do not affect the calculation directly. The result ranges from -100 to +100.

What is a good NPS for an MSP?

Industry benchmarks vary, but most IT service providers consider an NPS above +30 to be strong and above +50 to be excellent. A negative NPS (below zero) indicates more detractors than promoters and typically signals retention risk.

Why is our CSAT high but NPS low?

CSAT measures satisfaction with individual tickets. NPS measures overall loyalty and willingness to recommend. Clients can be satisfied with how tickets are resolved but still feel that the broader relationship lacks proactive communication, strategic value, or competitive pricing.

How often should we run NPS surveys?

Most MSPs run NPS surveys quarterly or biannually. Running them too frequently causes survey fatigue and lower response rates. The key is consistency: same cadence, same audience segmentation, so you can track trends over time.

What data sources does this report use?

This report uses SmileBack NPS survey responses (BI_SmileBack_Nps_Responses) for the 0-10 scores, SmileBack CSAT ticket reviews (BI_SmileBack_Reviews) for the positive/neutral/negative ratings, and Autotask PSA for client and ticket context.

Generate this report from your own data

Connect Proxuma Power BI to your PSA, RMM, and M365 environment, use an MCP-compatible AI to ask questions, and generate custom reports - in minutes, not days.

See more reports Get started