SlideShare una empresa de Scribd logo
1 de 18
Descargar para leer sin conexión
©IBM 2014
Spyros Kontogiorgis, PhD
with Brent Hodges, Tom Luin, Elmer Corbin
IBM Development Enterprise Transformation Initiative
2014 George Mason and IBM Symposium on Diverse Data Analytics Applications,
November 4th 2014.
Using analytics to capture the Voice of the
Customer and influence product development
2
Overview: Analytics for Customer Experience (CX)
 What is Quality and Customer Experience? Why measure it?
 How can we measure it? What metrics and methods can we use?
 Where can we find the underlying data and/or how can we create them?
 What types of analytics can we apply?
 What kind of insights can analytics produce?
 What new methods and tools does the future hold?
3
The evolution of Quality: From hard metrics to soft ones
 Definition: Quality is the extend to which products and services meet the
requirements of their users* .
 Quality has two aspects
– Quality of Design: The degree a product possesses an intended feature.
More features distinguish a product vs. the competition. (3 airbags are better than 1…)
– Quality of Conformance: The extend to which the product conforms to the
intent of the design.
The features perform as intended. (… provided they deploy when needed. )
 The supplier can control (track, improve) these aspects by instituting
metrics on
– the product itself
test results of compliance to specs (defects, rejects etc.).
– The production line and processes that create the product
setup time, down time, scrap, rework, completion time etc.
* “Introduction to Statistical Quality Control. 3rd Ed.”, by D.C. Montgomery (1996)
Some material adapted from “Measuring Customer Satisfaction, 2nd Ed.” by R.E. Hayes (1998)
4
Evolution of Quality (cont.): The case of services
 Supply-side metrics are “hard”, i.e. produced by an “objective” process.
 Experience has shown that hard metrics cannot capture (or forecast)
the success (adoption) of a product accurately, esp. of services, given
their unique characteristics*
Inseparability of production and consumption
High visibility to client. No buffer to hide mistakes or shortfalls.
Intangibility
Hard to test, measure, inventory, inspect and verify in advance of sale.
Perishability
Cannot store for later consumption. Must be right every time.
Heterogeneity
Performance varies with producers, delivery timeframe and customers expectation.
Hard to predict.
* Ghobadian et al, “Service quality: concepts and models”, Intl’ J. of Quality and Reliability Mgmt, 1997,
5
Evolution of Quality (cont.): Facets of soft metrics
 “Soft” metrics are introduced, to capture the subjective experience (and
perceptions, attitudes) of a user (= “market of one”) with an offering.
 We “measure” experience in a push (product-centric) or pull (user-centric) way:
*D. Garvin, “Managing Quality”, 1998
**J. Flanagan, “The critical incident technique”, Psychological Bulletin 1954, G. Latham et al., “Behavioral observation scales for performance
appraisal purposes, Personnel Psychology 1977.
Dimensions* Touchpoints / critical incidents**
Specific attributes of the product quality
deemed key
Points in time where a customer interacts with an
aspect of the product and forms an opinion
Performance (basic operating characteristics) Researching the product (web, marketing literature.)
Features (secondary characteristics) Contact with the sales force
Reliability (product will operate over time) Proposal quality
Conformance (product design and operation
meets standards)
Negotiation of features and price
Durability (product life) Delivery and installation
Serviceability (speed, competence, ease of repair) Technical support
Aesthetics (look-and-feel) Maintenance
Perception (image of company, reputation of
brand)
Trusted partner and source for future needs
6
“Soft” CX metrics are important: IBM Enterprise Transformation
mandates using them in all phases of the lifecycle of an offering
 Corporate Instruction ET 105 (of July 1, 2014) iterates that IBM
– views Quality through the eyes of our clients.
– is committed to continuously improve the overall quality experience for our
clients, which includes both
• their perception of product performance, as well as
• their interaction with IBM’s processes.
 To implement this commitment, ET 105 tasks
– Group Executives, to establish a Management system (with a focus on
Quality based on client experience) which will monitor the overall quality of
IBM offerings.
– Product Management, to set and implement aggressive client experience
Quality goals for their offerings.
– Design, to determine user experience focus areas.
7
To capture CX, we can use both push and pull methods
Feedback on customer experience can be captured by either of:
 A controlled (“push”) approach, such as surveys, with
– a crafted, officially-vetted questionnaire, directed at
– a segmented, often pre-qualified target audience, typically selected by
sampling methods, at
– pre-determined times (fixed dates, or after specific events, e.g. purchase,
registration, service completion), with
– responses given in numerical scales, and
– processed statistically.
 A free-form (“pull”) approach, such as data-mining opinions
– available in digital media (web sites, blogs, SMS, tweets etc.)
– unsolicited, and all available
– in text format (unstructured)
– scraped 24/7, and
– processed semantically.
8
Push practice: Capturing CX via trusted data surveys
 Continuously through the year, Market Development and Intelligence at
IBM (through affiliated vendors) field a series of surveys, which
 capture feedback from the leadership in customers’ organization (C-Suite and IT
managers who bought an offering recently), on an extensive range of interaction
facets with IBM
– 30+ questions (on delivery, pricing, support etc.), with replies scored in a 1-10 scale
(highest is 10), plus text comments to selected questions.
 are global (administered world-wide and across all brand groups. )
 are extensively used by the Marketing and Sales Organizations, and
 are data-maintained by a Business Intelligence Team, which syndicates them
through the enterprise.
 These surveys are the starting point in our analytics effort, with plans in
the future to expand into the CX analytics ecosystem…
9
It takes a diverse, sophisticated ecosystem to transmute CX from
raw data to actions informing business transformation
Capture Channels
• Purchase Surveys
• Board postings
• Chats
• Online Product Reviews
• Sensors (IoT)
Customer
Experience
Aggregators
• IT datamarts
• BI reports
• Automated On-
Page Recorders
Processors
• Embedded
Analytics Teams
• IBM Research
Brokers & Enablers
• Quality SMEs
• Leadership Teams
• Change Consultants
End Adopters
• Product Managers,
Designers & Developers
• Marketing and Sales
Analytics
Supply
Our team
Analytics
Demand
10
Which analytics for CX? Dashboards are an often indispensable
first step in displaying metrics and indicatorsComplexity
CompetitiveAdvantage
Ad hoc and Standard reporting
Query/drill down
Simulation
Forecasting
Statistical models
Optimization
Prescriptive
Descriptive
Predictive
Dashboards, although at just the first level of the analytics hierarchy, are a crucial
first step in generating business value from analytics.
 They provide values and visualizations of summary metrics (survey results and other CX
experience) – A “Close Encounter of the First Kind”* .
 They can be available on-demand 24/7, for rapid dissemination and adoption.
 They represent an advanced milestone of the journey travelled (and battles won) through
data capture, aggregation and processing.
 They can be used to trigger alerts, for special (outlier) events.
 They often provide a drill-down capability, used for elementary root-cause analysis.
Hypothesis testing
Dashboards
Stochastics
* ”Visual sightings of an unidentified flying object seemingly less than 500 feet away that show an appreciable angular extension and
considerable detail.” Hynek, Allen J. (1998) The UFO Experience: A Scientific Inquiry
11
However, to fully mine the CX experience, we need to go up the
analytics value chain, adding hypothesis testing
 Dashboards they may not be enough for decision making, since they
• show just one or two summary metrics, usually an average.
• show what happened in the past, which may lead to reactive (“fix it”) actions.
• show only what happened in one part of the business; the decision maker
must correlate with findings in other parts.
• may contain insights hidden; the decision maker must invest time to
discover them .
• provide no statistical valuation of variances (“is it a glitch or is it an issue?”)
 Thus, we must add higher-level analytics, such as hypothesis testing…
12
With our SME partners, we formulated and tested hypotheses
on client satisfaction
 The hypotheses come from suspected pain points, anecdotal info,
strategic and tactical goals and examination of patterns in the data.
Examples:
• Poorly-satisfied customers cannot be explained away as a random
variation (i.e. the distribution of satisfaction scores is skewed, not
normal.)
• The portion of poorly-satisfied customers is decreasing over time.
• Satisfaction levels have greater variability in service offerings, than in
hardware and software ones.
• Product defects is not one of the top issues contributing to negative
satisfaction.
• Clients at the exec-level would like IBM to provide guidance into future
technology trends proactively .
 While carrying out the analysis, we also sought to provide expressive
summaries (enhanced dashboard views, heat maps), to illustrate
findings in the data in a non-specialist language.
13
Example #1: A simple histogram by segment allows us to
understand changes in the overall satisfaction index over time
 Although the overall sat is high (60-70% of the responses in the high 20% end of the range -
skew), the poor sat zone is consistent over time, yet more than expected from a gaussian.
 The increase in the OK Sat zone explains the drop in the average Sat score over time.
Standard Dashboard View
Sat
score
Sat
zone
8-10 High
6-8 OK
0-6 Poor
Enhanced View with Segmentation
14
Example #2: Adding a simple linear trend and prediction
simplifies comparative satisfaction by brand.
 To gauge future behavior of poor sat, we added a 2-year-out linear trend to history.
 This extrapolation shows that poor sat with Red will continue to rise, while poor sat with
Yellow and Purple will decrease, and poor sat with Green will stay nearly flat.
15
Example #3: A heat map (geo drill down) reveals markets where
poor levels of sat require attention
 For the markets with the most
responses for A and B, we
segment their scores.
 We flag markets with
– High % of Poor sat as Red
– High % of OK sat as Yellow
– High % of High sat as Green.
 Prioritizing by # of responses
(proxy for business volume),
we identify markets in need of
immediate attention, and can
generate email alerts for.
Market Product A Product B
Poor
Sat
OK
High
Sat
#
responses
Poor
Sat
OK
High
Sat
#
responses
Segontium 3% 13% 84% 596 1% 11% 88% 224
Corinium 4% 34% 61% 174
Glevum 3% 15% 82% 114 19% 81% 16
Londinium 1% 26% 73% 74
Deva
Victrix
16% 37% 47% 70 100% 1
Eboracum 9% 26% 65% 46 13% 13% 75% 8
Mamucium 13% 35% 52% 23 33% 67% 6
Durnovaria 0% 19% 81% 21 100% 1
Lindinis 7% 36% 57% 14 60% 40% 5
Concangis 6% 44% 50% 18
Coria 6% 50% 44% 16 100% 2
Petuaria 7% 93% 14 100% 1
Sat
score
Sat
zone
0 to 6 Poor
6+ to 8 OK
8+ to 10 High
16
From Surveys to Social Listening: How to scale for Big Data?
 “The sage anticipates things that are difficult while they are easy, and
does things that would become great while they are small.
 All difficult things in the world are sure to arise from a previous state
in which they were easy, and all great things from one in which they
were small.
 Therefore the sage, while he never does what is great, is able on that
account to accomplish the greatest things.” (Dao De Jing)
As we establish and test the processes and checkpoints early, with
Medium-size Data, we build the foundations for handling Big Data.
Sponsor
Support
Process
Mapping
Stakeholder
interlock
Value
Creation
(PoC)
Data
Storage &
Ownership
Data
Sourcing
Data
Security
(ACL) Model
Validation
Data
Validation
Insight
Adoption
(Further)
Gap
Analysis
Deployment
Feedback
17
Continuous
Monitoring Continuous
Customer
Feedback and
Optimization
An Application Area: Leveraging Analytics for DevOps
Operate
Develop
and Test
Deploy
Steer
DevOps –
Continuous
Innovation
Product Process Data
Customer Feedback
Collect
Analyze
Inform
Launch
update
Breakdown of
silos between
Design,
Development
and Deployment
Release cycles
reduced to
weeks or days
(“daily dose”)
Products (apps)
are built with
basic analytics
instrumentation
Experiments in
customer
satisfaction are
easy to run
18
Into the future: Analytics by Cognitive Systems
 The Decision-Maker asks in natural language: “What are the main causes of
customer dissatisfaction? How may they affect orders in the pipeline?“
 The Cognitive System understands the question, produces possible answers
(hypotheses), collects supporting data, analyzes the evidence, computes the
confidence and delivers.
ARMONK, N.Y. - 09 Jan
2014: IBM today unveiled
three new Watson services
delivered over the cloud.
Watson Discovery Advisor is
designed to accelerate and
strengthen research and
development projects in
pharmaceuticals, publishing
and biotechnology. Watson
Analytics delivers visualized
Big Data insights, based on
questions posed in natural
language by any business
user. Watson Explorer helps
users across an enterprise
uncover and share data-driven
insights more easily, while
empowering organizations
launch Big Data initiatives
faster.Scenario

Más contenido relacionado

La actualidad más candente

Unit.1 . introduction to oprational research
Unit.1 . introduction to oprational researchUnit.1 . introduction to oprational research
Unit.1 . introduction to oprational researchDagnaygebawGoshme
 
Case study of LMD & HD trucks using Quality Function Deployment
Case study of LMD & HD trucks using Quality Function DeploymentCase study of LMD & HD trucks using Quality Function Deployment
Case study of LMD & HD trucks using Quality Function DeploymentIRJET Journal
 
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by Swink
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by SwinkTest Bank for Managing Operations Across the Supply Chain 2nd Edition by Swink
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by Swinkriven012
 
Basics of demand management for idiots
Basics of demand management for idiotsBasics of demand management for idiots
Basics of demand management for idiotsSambit T
 
Design and Market: Customer Requirements Market Survey
Design and Market: Customer Requirements Market SurveyDesign and Market: Customer Requirements Market Survey
Design and Market: Customer Requirements Market SurveyNaseel Ibnu Azeez
 
Product Design & Development
Product Design & DevelopmentProduct Design & Development
Product Design & DevelopmentQRCE
 
104vendor Appraisal
104vendor Appraisal104vendor Appraisal
104vendor Appraisalbtecexpert
 
Requirement prioritization
Requirement prioritizationRequirement prioritization
Requirement prioritizationAbdul Basit
 
Product Life Cycle Management
Product Life Cycle ManagementProduct Life Cycle Management
Product Life Cycle ManagementAnand Subramaniam
 
Lect 2 project screening & selection
Lect 2 project screening & selection Lect 2 project screening & selection
Lect 2 project screening & selection Muhammad Bilal
 
Deep Dive Into Demand Planning
Deep Dive Into Demand PlanningDeep Dive Into Demand Planning
Deep Dive Into Demand Planningjkirchner
 
Smart Sim Selector: A Software for Simulation Software Selection
Smart Sim Selector: A Software for Simulation Software SelectionSmart Sim Selector: A Software for Simulation Software Selection
Smart Sim Selector: A Software for Simulation Software SelectionCSCJournals
 
Analytics for Process Excellence
Analytics for Process ExcellenceAnalytics for Process Excellence
Analytics for Process ExcellenceDenis Gagné
 
Product design and development
Product design and developmentProduct design and development
Product design and developmentVishal Kachhdiya
 
Software Engineering Economics Life Cycle.
Software Engineering Economics  Life Cycle.Software Engineering Economics  Life Cycle.
Software Engineering Economics Life Cycle.Sulman Ahmed
 
Product and service design
Product and service designProduct and service design
Product and service designGrace Falcis
 

La actualidad más candente (20)

3rd unit
3rd unit3rd unit
3rd unit
 
Unit.1 . introduction to oprational research
Unit.1 . introduction to oprational researchUnit.1 . introduction to oprational research
Unit.1 . introduction to oprational research
 
Case study of LMD & HD trucks using Quality Function Deployment
Case study of LMD & HD trucks using Quality Function DeploymentCase study of LMD & HD trucks using Quality Function Deployment
Case study of LMD & HD trucks using Quality Function Deployment
 
BPSim Briefing
BPSim BriefingBPSim Briefing
BPSim Briefing
 
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by Swink
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by SwinkTest Bank for Managing Operations Across the Supply Chain 2nd Edition by Swink
Test Bank for Managing Operations Across the Supply Chain 2nd Edition by Swink
 
Basics of demand management for idiots
Basics of demand management for idiotsBasics of demand management for idiots
Basics of demand management for idiots
 
Design and Market: Customer Requirements Market Survey
Design and Market: Customer Requirements Market SurveyDesign and Market: Customer Requirements Market Survey
Design and Market: Customer Requirements Market Survey
 
Product Design & Development
Product Design & DevelopmentProduct Design & Development
Product Design & Development
 
QFD For Pizza
QFD For PizzaQFD For Pizza
QFD For Pizza
 
104vendor Appraisal
104vendor Appraisal104vendor Appraisal
104vendor Appraisal
 
Requirement prioritization
Requirement prioritizationRequirement prioritization
Requirement prioritization
 
Product Life Cycle Management
Product Life Cycle ManagementProduct Life Cycle Management
Product Life Cycle Management
 
VALUE ANALYSIS
VALUE ANALYSISVALUE ANALYSIS
VALUE ANALYSIS
 
Lect 2 project screening & selection
Lect 2 project screening & selection Lect 2 project screening & selection
Lect 2 project screening & selection
 
Deep Dive Into Demand Planning
Deep Dive Into Demand PlanningDeep Dive Into Demand Planning
Deep Dive Into Demand Planning
 
Smart Sim Selector: A Software for Simulation Software Selection
Smart Sim Selector: A Software for Simulation Software SelectionSmart Sim Selector: A Software for Simulation Software Selection
Smart Sim Selector: A Software for Simulation Software Selection
 
Analytics for Process Excellence
Analytics for Process ExcellenceAnalytics for Process Excellence
Analytics for Process Excellence
 
Product design and development
Product design and developmentProduct design and development
Product design and development
 
Software Engineering Economics Life Cycle.
Software Engineering Economics  Life Cycle.Software Engineering Economics  Life Cycle.
Software Engineering Economics Life Cycle.
 
Product and service design
Product and service designProduct and service design
Product and service design
 

Similar a Analytics for the Voice of the Customer - SK

Software Productivity Framework
Software Productivity Framework Software Productivity Framework
Software Productivity Framework Zinnov
 
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019How to Carry Out a Design Due Diligence? | Seattle Interactive 2019
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019Seattle Interactive Conference
 
Operationalizing Customer Analytics with Azure and Power BI
Operationalizing Customer Analytics with Azure and Power BIOperationalizing Customer Analytics with Azure and Power BI
Operationalizing Customer Analytics with Azure and Power BICCG
 
Six Sigma Yellow Belt Training
Six Sigma Yellow Belt TrainingSix Sigma Yellow Belt Training
Six Sigma Yellow Belt TrainingInvensis Learning
 
Innovate iq digital transformation by webonise
Innovate iq   digital transformation by weboniseInnovate iq   digital transformation by webonise
Innovate iq digital transformation by weboniseBhuvan Khanna
 
Agility Accelerator
Agility AcceleratorAgility Accelerator
Agility AcceleratorCraig Smith
 
Customer Experience Buyer's Guide
Customer Experience Buyer's GuideCustomer Experience Buyer's Guide
Customer Experience Buyer's GuideInMoment
 
ICF-related data: the new frontier of individualised, predictive healthcare ]...
ICF-related data: the new frontier of individualised, predictive healthcare ]...ICF-related data: the new frontier of individualised, predictive healthcare ]...
ICF-related data: the new frontier of individualised, predictive healthcare ]...ICF Education
 
Marketing Analytics at IBM - export
Marketing Analytics at IBM - exportMarketing Analytics at IBM - export
Marketing Analytics at IBM - exportSpyros Kontogiorgis
 
Psychic insights with survey analytics
Psychic insights with survey analyticsPsychic insights with survey analytics
Psychic insights with survey analyticsAmit Pishe
 
Analytics in manufacturing
Analytics in manufacturingAnalytics in manufacturing
Analytics in manufacturingSaurav Kumar
 
HWZ-Darden Konferenz: Building a Sustainable Analytics Orientation
HWZ-Darden Konferenz: Building a Sustainable Analytics OrientationHWZ-Darden Konferenz: Building a Sustainable Analytics Orientation
HWZ-Darden Konferenz: Building a Sustainable Analytics OrientationHWZ Hochschule für Wirtschaft
 
Chainsaw Conjoint
Chainsaw ConjointChainsaw Conjoint
Chainsaw ConjointQuestionPro
 
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...JK Tech
 
Topic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxTopic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxCallplanetsDeveloper
 
OPERATIONS MANAGEMENT Chapter 1 5
OPERATIONS MANAGEMENT Chapter 1 5OPERATIONS MANAGEMENT Chapter 1 5
OPERATIONS MANAGEMENT Chapter 1 5yared ayele
 
Risk Driven Testing
Risk Driven TestingRisk Driven Testing
Risk Driven TestingJorge Boria
 
Job-to-be-done theory to practice : Ch4 Process
Job-to-be-done theory to practice : Ch4 ProcessJob-to-be-done theory to practice : Ch4 Process
Job-to-be-done theory to practice : Ch4 ProcessPRADA Hsiung
 

Similar a Analytics for the Voice of the Customer - SK (20)

Software Productivity Framework
Software Productivity Framework Software Productivity Framework
Software Productivity Framework
 
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019How to Carry Out a Design Due Diligence? | Seattle Interactive 2019
How to Carry Out a Design Due Diligence? | Seattle Interactive 2019
 
Operationalizing Customer Analytics with Azure and Power BI
Operationalizing Customer Analytics with Azure and Power BIOperationalizing Customer Analytics with Azure and Power BI
Operationalizing Customer Analytics with Azure and Power BI
 
Six Sigma Yellow Belt Training
Six Sigma Yellow Belt TrainingSix Sigma Yellow Belt Training
Six Sigma Yellow Belt Training
 
Innovate iq digital transformation by webonise
Innovate iq   digital transformation by weboniseInnovate iq   digital transformation by webonise
Innovate iq digital transformation by webonise
 
Agility Accelerator
Agility AcceleratorAgility Accelerator
Agility Accelerator
 
Customer Experience Buyer's Guide
Customer Experience Buyer's GuideCustomer Experience Buyer's Guide
Customer Experience Buyer's Guide
 
ICF-related data: the new frontier of individualised, predictive healthcare ]...
ICF-related data: the new frontier of individualised, predictive healthcare ]...ICF-related data: the new frontier of individualised, predictive healthcare ]...
ICF-related data: the new frontier of individualised, predictive healthcare ]...
 
Marketing Analytics at IBM - export
Marketing Analytics at IBM - exportMarketing Analytics at IBM - export
Marketing Analytics at IBM - export
 
Advanced Analytics
Advanced AnalyticsAdvanced Analytics
Advanced Analytics
 
Psychic insights with survey analytics
Psychic insights with survey analyticsPsychic insights with survey analytics
Psychic insights with survey analytics
 
CORPORATE CRITERION
CORPORATE CRITERIONCORPORATE CRITERION
CORPORATE CRITERION
 
Analytics in manufacturing
Analytics in manufacturingAnalytics in manufacturing
Analytics in manufacturing
 
HWZ-Darden Konferenz: Building a Sustainable Analytics Orientation
HWZ-Darden Konferenz: Building a Sustainable Analytics OrientationHWZ-Darden Konferenz: Building a Sustainable Analytics Orientation
HWZ-Darden Konferenz: Building a Sustainable Analytics Orientation
 
Chainsaw Conjoint
Chainsaw ConjointChainsaw Conjoint
Chainsaw Conjoint
 
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...
[AI Webinar Series P1] - How Advanced Text Analytics Can Increase the Operati...
 
Topic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptxTopic 12 Miscellaneous Concepts in IT.pptx
Topic 12 Miscellaneous Concepts in IT.pptx
 
OPERATIONS MANAGEMENT Chapter 1 5
OPERATIONS MANAGEMENT Chapter 1 5OPERATIONS MANAGEMENT Chapter 1 5
OPERATIONS MANAGEMENT Chapter 1 5
 
Risk Driven Testing
Risk Driven TestingRisk Driven Testing
Risk Driven Testing
 
Job-to-be-done theory to practice : Ch4 Process
Job-to-be-done theory to practice : Ch4 ProcessJob-to-be-done theory to practice : Ch4 Process
Job-to-be-done theory to practice : Ch4 Process
 

Analytics for the Voice of the Customer - SK

  • 1. ©IBM 2014 Spyros Kontogiorgis, PhD with Brent Hodges, Tom Luin, Elmer Corbin IBM Development Enterprise Transformation Initiative 2014 George Mason and IBM Symposium on Diverse Data Analytics Applications, November 4th 2014. Using analytics to capture the Voice of the Customer and influence product development
  • 2. 2 Overview: Analytics for Customer Experience (CX)  What is Quality and Customer Experience? Why measure it?  How can we measure it? What metrics and methods can we use?  Where can we find the underlying data and/or how can we create them?  What types of analytics can we apply?  What kind of insights can analytics produce?  What new methods and tools does the future hold?
  • 3. 3 The evolution of Quality: From hard metrics to soft ones  Definition: Quality is the extend to which products and services meet the requirements of their users* .  Quality has two aspects – Quality of Design: The degree a product possesses an intended feature. More features distinguish a product vs. the competition. (3 airbags are better than 1…) – Quality of Conformance: The extend to which the product conforms to the intent of the design. The features perform as intended. (… provided they deploy when needed. )  The supplier can control (track, improve) these aspects by instituting metrics on – the product itself test results of compliance to specs (defects, rejects etc.). – The production line and processes that create the product setup time, down time, scrap, rework, completion time etc. * “Introduction to Statistical Quality Control. 3rd Ed.”, by D.C. Montgomery (1996) Some material adapted from “Measuring Customer Satisfaction, 2nd Ed.” by R.E. Hayes (1998)
  • 4. 4 Evolution of Quality (cont.): The case of services  Supply-side metrics are “hard”, i.e. produced by an “objective” process.  Experience has shown that hard metrics cannot capture (or forecast) the success (adoption) of a product accurately, esp. of services, given their unique characteristics* Inseparability of production and consumption High visibility to client. No buffer to hide mistakes or shortfalls. Intangibility Hard to test, measure, inventory, inspect and verify in advance of sale. Perishability Cannot store for later consumption. Must be right every time. Heterogeneity Performance varies with producers, delivery timeframe and customers expectation. Hard to predict. * Ghobadian et al, “Service quality: concepts and models”, Intl’ J. of Quality and Reliability Mgmt, 1997,
  • 5. 5 Evolution of Quality (cont.): Facets of soft metrics  “Soft” metrics are introduced, to capture the subjective experience (and perceptions, attitudes) of a user (= “market of one”) with an offering.  We “measure” experience in a push (product-centric) or pull (user-centric) way: *D. Garvin, “Managing Quality”, 1998 **J. Flanagan, “The critical incident technique”, Psychological Bulletin 1954, G. Latham et al., “Behavioral observation scales for performance appraisal purposes, Personnel Psychology 1977. Dimensions* Touchpoints / critical incidents** Specific attributes of the product quality deemed key Points in time where a customer interacts with an aspect of the product and forms an opinion Performance (basic operating characteristics) Researching the product (web, marketing literature.) Features (secondary characteristics) Contact with the sales force Reliability (product will operate over time) Proposal quality Conformance (product design and operation meets standards) Negotiation of features and price Durability (product life) Delivery and installation Serviceability (speed, competence, ease of repair) Technical support Aesthetics (look-and-feel) Maintenance Perception (image of company, reputation of brand) Trusted partner and source for future needs
  • 6. 6 “Soft” CX metrics are important: IBM Enterprise Transformation mandates using them in all phases of the lifecycle of an offering  Corporate Instruction ET 105 (of July 1, 2014) iterates that IBM – views Quality through the eyes of our clients. – is committed to continuously improve the overall quality experience for our clients, which includes both • their perception of product performance, as well as • their interaction with IBM’s processes.  To implement this commitment, ET 105 tasks – Group Executives, to establish a Management system (with a focus on Quality based on client experience) which will monitor the overall quality of IBM offerings. – Product Management, to set and implement aggressive client experience Quality goals for their offerings. – Design, to determine user experience focus areas.
  • 7. 7 To capture CX, we can use both push and pull methods Feedback on customer experience can be captured by either of:  A controlled (“push”) approach, such as surveys, with – a crafted, officially-vetted questionnaire, directed at – a segmented, often pre-qualified target audience, typically selected by sampling methods, at – pre-determined times (fixed dates, or after specific events, e.g. purchase, registration, service completion), with – responses given in numerical scales, and – processed statistically.  A free-form (“pull”) approach, such as data-mining opinions – available in digital media (web sites, blogs, SMS, tweets etc.) – unsolicited, and all available – in text format (unstructured) – scraped 24/7, and – processed semantically.
  • 8. 8 Push practice: Capturing CX via trusted data surveys  Continuously through the year, Market Development and Intelligence at IBM (through affiliated vendors) field a series of surveys, which  capture feedback from the leadership in customers’ organization (C-Suite and IT managers who bought an offering recently), on an extensive range of interaction facets with IBM – 30+ questions (on delivery, pricing, support etc.), with replies scored in a 1-10 scale (highest is 10), plus text comments to selected questions.  are global (administered world-wide and across all brand groups. )  are extensively used by the Marketing and Sales Organizations, and  are data-maintained by a Business Intelligence Team, which syndicates them through the enterprise.  These surveys are the starting point in our analytics effort, with plans in the future to expand into the CX analytics ecosystem…
  • 9. 9 It takes a diverse, sophisticated ecosystem to transmute CX from raw data to actions informing business transformation Capture Channels • Purchase Surveys • Board postings • Chats • Online Product Reviews • Sensors (IoT) Customer Experience Aggregators • IT datamarts • BI reports • Automated On- Page Recorders Processors • Embedded Analytics Teams • IBM Research Brokers & Enablers • Quality SMEs • Leadership Teams • Change Consultants End Adopters • Product Managers, Designers & Developers • Marketing and Sales Analytics Supply Our team Analytics Demand
  • 10. 10 Which analytics for CX? Dashboards are an often indispensable first step in displaying metrics and indicatorsComplexity CompetitiveAdvantage Ad hoc and Standard reporting Query/drill down Simulation Forecasting Statistical models Optimization Prescriptive Descriptive Predictive Dashboards, although at just the first level of the analytics hierarchy, are a crucial first step in generating business value from analytics.  They provide values and visualizations of summary metrics (survey results and other CX experience) – A “Close Encounter of the First Kind”* .  They can be available on-demand 24/7, for rapid dissemination and adoption.  They represent an advanced milestone of the journey travelled (and battles won) through data capture, aggregation and processing.  They can be used to trigger alerts, for special (outlier) events.  They often provide a drill-down capability, used for elementary root-cause analysis. Hypothesis testing Dashboards Stochastics * ”Visual sightings of an unidentified flying object seemingly less than 500 feet away that show an appreciable angular extension and considerable detail.” Hynek, Allen J. (1998) The UFO Experience: A Scientific Inquiry
  • 11. 11 However, to fully mine the CX experience, we need to go up the analytics value chain, adding hypothesis testing  Dashboards they may not be enough for decision making, since they • show just one or two summary metrics, usually an average. • show what happened in the past, which may lead to reactive (“fix it”) actions. • show only what happened in one part of the business; the decision maker must correlate with findings in other parts. • may contain insights hidden; the decision maker must invest time to discover them . • provide no statistical valuation of variances (“is it a glitch or is it an issue?”)  Thus, we must add higher-level analytics, such as hypothesis testing…
  • 12. 12 With our SME partners, we formulated and tested hypotheses on client satisfaction  The hypotheses come from suspected pain points, anecdotal info, strategic and tactical goals and examination of patterns in the data. Examples: • Poorly-satisfied customers cannot be explained away as a random variation (i.e. the distribution of satisfaction scores is skewed, not normal.) • The portion of poorly-satisfied customers is decreasing over time. • Satisfaction levels have greater variability in service offerings, than in hardware and software ones. • Product defects is not one of the top issues contributing to negative satisfaction. • Clients at the exec-level would like IBM to provide guidance into future technology trends proactively .  While carrying out the analysis, we also sought to provide expressive summaries (enhanced dashboard views, heat maps), to illustrate findings in the data in a non-specialist language.
  • 13. 13 Example #1: A simple histogram by segment allows us to understand changes in the overall satisfaction index over time  Although the overall sat is high (60-70% of the responses in the high 20% end of the range - skew), the poor sat zone is consistent over time, yet more than expected from a gaussian.  The increase in the OK Sat zone explains the drop in the average Sat score over time. Standard Dashboard View Sat score Sat zone 8-10 High 6-8 OK 0-6 Poor Enhanced View with Segmentation
  • 14. 14 Example #2: Adding a simple linear trend and prediction simplifies comparative satisfaction by brand.  To gauge future behavior of poor sat, we added a 2-year-out linear trend to history.  This extrapolation shows that poor sat with Red will continue to rise, while poor sat with Yellow and Purple will decrease, and poor sat with Green will stay nearly flat.
  • 15. 15 Example #3: A heat map (geo drill down) reveals markets where poor levels of sat require attention  For the markets with the most responses for A and B, we segment their scores.  We flag markets with – High % of Poor sat as Red – High % of OK sat as Yellow – High % of High sat as Green.  Prioritizing by # of responses (proxy for business volume), we identify markets in need of immediate attention, and can generate email alerts for. Market Product A Product B Poor Sat OK High Sat # responses Poor Sat OK High Sat # responses Segontium 3% 13% 84% 596 1% 11% 88% 224 Corinium 4% 34% 61% 174 Glevum 3% 15% 82% 114 19% 81% 16 Londinium 1% 26% 73% 74 Deva Victrix 16% 37% 47% 70 100% 1 Eboracum 9% 26% 65% 46 13% 13% 75% 8 Mamucium 13% 35% 52% 23 33% 67% 6 Durnovaria 0% 19% 81% 21 100% 1 Lindinis 7% 36% 57% 14 60% 40% 5 Concangis 6% 44% 50% 18 Coria 6% 50% 44% 16 100% 2 Petuaria 7% 93% 14 100% 1 Sat score Sat zone 0 to 6 Poor 6+ to 8 OK 8+ to 10 High
  • 16. 16 From Surveys to Social Listening: How to scale for Big Data?  “The sage anticipates things that are difficult while they are easy, and does things that would become great while they are small.  All difficult things in the world are sure to arise from a previous state in which they were easy, and all great things from one in which they were small.  Therefore the sage, while he never does what is great, is able on that account to accomplish the greatest things.” (Dao De Jing) As we establish and test the processes and checkpoints early, with Medium-size Data, we build the foundations for handling Big Data. Sponsor Support Process Mapping Stakeholder interlock Value Creation (PoC) Data Storage & Ownership Data Sourcing Data Security (ACL) Model Validation Data Validation Insight Adoption (Further) Gap Analysis Deployment Feedback
  • 17. 17 Continuous Monitoring Continuous Customer Feedback and Optimization An Application Area: Leveraging Analytics for DevOps Operate Develop and Test Deploy Steer DevOps – Continuous Innovation Product Process Data Customer Feedback Collect Analyze Inform Launch update Breakdown of silos between Design, Development and Deployment Release cycles reduced to weeks or days (“daily dose”) Products (apps) are built with basic analytics instrumentation Experiments in customer satisfaction are easy to run
  • 18. 18 Into the future: Analytics by Cognitive Systems  The Decision-Maker asks in natural language: “What are the main causes of customer dissatisfaction? How may they affect orders in the pipeline?“  The Cognitive System understands the question, produces possible answers (hypotheses), collects supporting data, analyzes the evidence, computes the confidence and delivers. ARMONK, N.Y. - 09 Jan 2014: IBM today unveiled three new Watson services delivered over the cloud. Watson Discovery Advisor is designed to accelerate and strengthen research and development projects in pharmaceuticals, publishing and biotechnology. Watson Analytics delivers visualized Big Data insights, based on questions posed in natural language by any business user. Watson Explorer helps users across an enterprise uncover and share data-driven insights more easily, while empowering organizations launch Big Data initiatives faster.Scenario