CLIRS: Customer Loyalty Improvement Recommender System.

Slides:



Advertisements
Similar presentations
© Megaputer intelligence, Inc. Your Knowledge Partner Survey Analysis using PolyAnalyst TM.
Advertisements

Data Mining Tools Overview Business Intelligence for Managers.
Introducing… Powered by 1 An immediate, systematic, customer follow-up and complaint resolution program designed for dealers by dealers Powered by.
Title Course opinion mining methodology for knowledge discovery, based on web social media Authors Sotirios Kontogiannis Ioannis Kazanidis Stavros Valsamidis.
April 22, Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Doerre, Peter Gerstl, Roland Seiffert IBM Germany, August 1999 Presenter:
Data Mining By Archana Ketkar.
Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Huimin Ye.
Text Mining: Finding Nuggets in Mountains of Textual Data Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Drew DeHaas.
Data Mining Techniques
FALL 2012 DSCI5240 Graduate Presentation By Xxxxxxx.
Data Mining. 2 Models Created by Data Mining Linear Equations Rules Clusters Graphs Tree Structures Recurrent Patterns.
CIS 9002 Kannan Mohan Department of CIS Zicklin School of Business, Baruch College.
© Copyright 2010 Hewlett-Packard Development Company, L.P. 1 Swati Saxena Senior Research Manager August RESEARCH AND ANALYTICS: AN SMB CASE STUDY.
Georgia Credit Union Idea Institute 10 Tammy Gallegos SVP Marketing Communications & Service Quality August 2015.
An-Najah National University Faculty Of Engineering Industrial Engineering Department Implementation Of Quality Function Deployment On Engineering Faculty.
© 2009 Cisco Systems, Inc. All rights reserved.Cisco Confidential Cisco SBCS Solution for Auto Repair Accelerate the productivity of your shop.
6.1 © 2010 by Prentice Hall 6 Chapter Foundations of Business Intelligence: Databases and Information Management.
3-1 Data Mining Kelby Lee. 3-2 Overview ¨ Transaction Database ¨ What is Data Mining ¨ Data Mining Primitives ¨ Data Mining Objectives ¨ Predictive Modeling.
INFORMATION SYSTEMS FOR MANAGEMENT. Agenda Information system project Organization analysis.
Customer Preferences for Metering and Connectivity Metering Americas 2004 San Diego, CA March 24-26, 2004 Lynn Fryer Stein Primen.
With Doug Atkins Getting Data Out of FASTER: Tips for the New & Experienced.
Data Mining BY JEMINI ISLAM. Data Mining Outline: What is data mining? Why use data mining? How does data mining work The process of data mining Tools.
Why BI….? Most companies collect a large amount of data from their business operations. To keep track of that information, a business and would need to.
Text Information Management ChengXiang Zhai, Tao Tao, Xuehua Shen, Hui Fang, Azadeh Shakery, Jing Jiang.
Jochen Dijrre, Peter Gerstl, Roland Seiffert Presented by Shamil Mustafayev 04/16/
RESEARCH MOTHODOLOGY SZRZ6014 Dr. Farzana Kabir Ahmad Taqiyah Khadijah Ghazali (814537) SENTIMENT ANALYSIS FOR VOICE OF THE CUSTOMER.
MarketsandMarkets Presents MarketsandMarkets Presents Customer Experience Management Market Expected to Reach $6.61 Billion by 2017 Customer Experience.
Early Warning and Real Time Anomaly Detection System © 2015 Blueocean Market Intelligence1.
The Power of Using Artificial Intelligence
The Premier Test Lab Management Software
Take the Guesswork Out of Calculating Your Project Cost
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
What is CRM ? Customer Relationship Management (CRM) is an information industry term for methodologies, software, and usually Internet capabilities that.
Customer Support.
Hierarchical Recommender System for Improving NPS.
Can we build recommender system for artwork evaluation?
Chris Turner Business Advantage
NPS-Based Recommender System for Increasing Business Revenue.
Future-oriented Benchmarking Through Social Media Analysis
Future-Oriented Benchmarking through Social Media Analysis
Insights driven Customer Experience
ENHANCING MANAGEMENT DECISION MAKING
Applications of Data Mining in Software Engineering
ServiceNow Assessments
From Knowledge Discovery to Customer Attrition
My Learning from SSGA!.
What issues keep you up at night?
Employee Engagement 2018 Example of analysis report
CCI, UNC-Charlotte User-friendly NPS-based recommender system for driving business revenue. presented by Zbigniew W. Ras Research sponsored by WI’17, Leipzig,
Supporting End-User Access
Data Quality in the BI Life Cycle
Course Lab Introduction to IBM Watson Analytics
Janet LeBlanc Director, Canada Post
HFFA Membership Study Results Wave 2
AGMLAB Information Technologies
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
PolyAnalyst Web Report Training
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Customer Satisfaction Survey Report For
Megaputer Intelligence
Roadside Summary Report– November 2015
PolyAnalyst Web Report Training
Quick Housekeeping Questions Webinar Recording + Slides Webinar Survey
Business Analytics Professional Services
Discovering Companies we Know
Helping Agents Grow Based on Customer Insights
Chapter 13: Project Stakeholder Management
Ontology-Enhanced Aspect-Based Sentiment Analysis
AI Builder for Power Platform
Presentation transcript:

CLIRS: Customer Loyalty Improvement Recommender System. Built for the Daniel Group (2013-2018) presented by Zbigniew W. Ras Research sponsored by

Project Team KDD Lab Members Kasia Tarnowska (San Jose State University) Pauline Brunet (Paris Tech., France) Jieyan Kuang (Bank of America) Z.W.Ras Industry Collaborators Yuehua Duan PhD Student Lynn Daniel Founder & CEO Doug Fowler CHIEF OPERATING OFFICER

Consulting Company CUSTOMERS Client 1 Client 2 Client 3 Client 4 shops What we had to do Build Recommender System for each client (34 clients) helping to increase its revenue Client 1 Client 2 Client 3 Client 4 shops shops shops Build Personalized Recommender System for each shop helping to increase its revenue Services (heavy equipment repair) Parts CUSTOMERS

NPS is correlated with company revenue growth Net Promoter Score (NPS) – today’s standard for measuring customer loyalty Promoter - {9,10} Passive – {8,7} Detractor – {1,2,3…,6}

GET KEEP GROW CAC - Average expense of gaining a single customer LTV – Customer Lifetime Value

NPS rating for all clients

NPS rating for all shops

CLIRS Our Software

Temper: https://www.temper.io/

Temper Collecting Customers Data Limited Analytics

Clarabridge CX Analytics [http://www.clarabridge.com/] Key Features: Deep Natural Language Processing (NLP) - Accurately analyze mountains of unstructured data to zero in on aspects of business that drive customers’ dissatisfaction. Linguistic categorization - Intelligently groups customer comments into topic buckets for smarter analysis. Emotion detection - Decipher the emotional state of customers based on the tone of the feedback. Context-sensitive sentiment analysis - Understands the intensity of feelings expressed by using deep Natural Language Processing, adjusting for industry and source-specific nuances. Advanced discovery tools - Get to the core of customer issues using single click root cause analysis; unearth previously unknown topics using fully automated topic detection engine. Speech analytics – Capture customer’s voice, literally, by analyzing call recordings in the same platform as all your other data sources. Does not discover actionable knowledge (recommendations) which guarantee NPS improvement, especially how large it is TATA Company: collects not only customers text data but also voice & video

CLIRS Our System

Current samplings: about 300,000 records per year, 2010-2018

Decision Table built from customers survey M/F Phone Bench1 Bench2 Bench3 Comments Promoter Status 2011 M 3 9 5 text1 Promoter 2012 8 10 text2 2013 F text3 Detractor 2014 7 text4 Passive New attributes can be derived Text Mining & Sentiment Mining can be used to build new attributes

PLAN FOR TODAY’s PRESENTATION CLIRS - Customer Loyalty Improvement Recommender System Definitions/Concepts needed to understand how CLIRS is built & works: Reducts in Decision Systems Action Rules and Meta-Actions Decision Tables & Their Semantic Similarity

? Decision System - Granules {x3,x4} Promoter Bench1 Bench2 Promoter Status x1 a1 b2 promoter x2 a1 b1 passive x3 a2 b3 passive x4 a2 b3 promoter x5 a3 b4 passive x6 a1 b4 promoter x7 a3 b4 passive Decision Granules: {x1,x4,x6}, {x2,x4,x6} Classification Granules: Bench1: {x1,x2,x6}, {x3,x4}, {x5,x7} Bench2: {x1},{x2},{x3,x4},{x5,x6,x7} Smallest Granules: {x1},{x2},{x3,x4},{x5,x7},{x6} Passive {x2}, {x5,x7} Passive {x3,x4} Informal Definition: Reduct – smallest subset of classification attributes which preserves distribution of objects in Figure 1 {x1}, {x6} Promoter/Passive Promoter ?

Decision System & Reducts (Rough Sets) Reduct1 = {Muscle-pain,Temp.} Reduct2 = {Headache, Temp.} We are looking for rules describing Flu in terms of Headache, Muscle Pain, Temp.

ACTION RULES & META ACTIONS Action Rules Miner: Action4ft-Miner module in LISP Miner http://lispminer.vse.cz/procedures/index.php?system=Ac4ftMiner [developed by Jan Rauch (Univ. of Economics, Prague, Czech Republic] Meta Action Miners: Only Domain Specific (parts of domain specific software) Triggers of Action Rules

Smaller the distance between tables, more similar they are. R1 R3 R2 Dk Dn

Coming back to our customers feedback dataset Rough Set REDUCTS have been used to identify features having the strongest impact on Promoter Benchmark All Overall Satisfaction 35 Benchmark All Likelihood to be Repeat Customer 34 Benchmark All Dealer Communication 33 Benchmark Service Repair Completed Correctly 32 Benchmark Referral Behavior 31 Benchmark Service Final Invoice Matched Expectations Benchmark Ease of Contact 30 Benchmark All Does Customer have Future Needs 28 Benchmark Service Tech Promised in Expected Timeframe 26 Benchmark Service Repair Completed When Promised Benchmark Service Timeliness of Invoice 25 Benchmark Service Appointment Availability 24 Benchmark Service Tech Equipped to do Job 23 Benchmark All Contact Status of Future Needs 22 Benchmark Service Tech Arrived When Promised 21 Benchmark All Has Issue Been Resolved 19 Benchmark All Contact Status of Issue 17 Benchmark Service Technician Communication 6 Benchmark Service Contact Preference 3 Benchmark CB Call answered promptly 1 Benchmark Service Received Quote for Repair Benchmark CB Auto attendant answered by correct department Benchmark Service Call Answered Quickly Benchmark All Marketing Permission Randomly chosen customers are asked to complete Questionnaire. It has questions concerning personal data + 30 benchmarks To compute NPS we calculate average score of selected benchmarks for all customers. Knowing the number of promoters and detractors we know NPS.

RECOMMENDER SYSTEM CLIRS based on semantic extension of a client dataset created by adding datasets of other semantically similar clients

Extracting Meta-Actions DATASET Guided Folksonomy & Sentiment Analysis Meta-Actions Extracting Comments C1 C2 C3 C4 C2,C4 Knowledgeable Staff Customer Bench1 Bench2 Bench3 Comments Prom_Stat Cust1 high med high C1, C2 Prom C1,C3 Friendly Staff Cust2 med med med C2, C3 Pass C1 refers to Bench2 C3 refers to Bench1 C2,C4 refers to Bench3 Cust3 C4, C3 Benchmark Values ={low, medium, high} Extracted Rules (example) R1=[ (Bench1=high)&(Bench2=med) …………......  (Prom_Stat = Prom)] sup(R1)={Cust1,..}; Let’s say Confidence=90% R2=[ (Bench1=med)& ……………. (Bench3=med)  (Prom_Stat = Pass)] sup(R2)={Cust2,..}; Let’s say Confidence=95% Action Rules R= [(Bench1, med ->high)&(Bench2=med)  (Prom_Stat, Pass -> Prom)] Sup(R)={Cust2,..}; Confidence=90%*95%=85.5%. Recommendation: Staff has to be more friendly (R will be activated)

Recommender System

User-Friendly Interface

Recommender System

DATASET c3 Step 2: Guided folksonomy & Sentiment analysis of comments Names of Clusters Step 1: Extracting Comments C2,C4 Attr3 C1 C2 C3 C4 Painting Attr1 Attr2 Attr3 Comments Price p1 neutral positive C1, C2 [m1,m2] C1 Attr2 p2 positive positive C2, C3 [m3,m4] C1 - neutral C3 - positive C2,- positive C4 - negative p3 positive negative C4, C3 [m1,m2] c3 Attr1 Step 3: Attr1, Attr2, Attr3 – Names of new attributes & their values added to the dataset