30
Galit Shmuéli SRITNE Chaired Professor of Data Analytics Predicting, Explaining and the Business Analytics Toolkit

NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

  • Upload
    nasscom

  • View
    5.325

  • Download
    4

Embed Size (px)

DESCRIPTION

Presentations by Prof. Galit Shmuéli, SRITNE Chaired Professor of Data Analytics, ISB at NASSCOM Big Data and Analytics Summit 2014.

Citation preview

Page 1: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Galit ShmuéliSRITNE Chaired

Professor of Data Analytics

Predicting, Explaining and the Business Analytics Toolkit

Page 2: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Business Intelligence

Traditional: Describe the past

State-of-the-Art: Describe the present

Business Analytics

Predictive Analytics: Predict future of individual records

Explanatory Analytics: Explain cause-effect of “average record”

(overall effect)

Page 3: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Today’s Talk

1. Predictive Analytics: The process & applications

2. Prediction is not explanation

3. The Explanatory Analytics toolkit

Page 4: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Will the customer pay?

What causes non-payment?

Page 5: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli
Page 6: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Past Present Future

Case Studies

Overall Behaviour

“Presonalized” Behaviour

Page 7: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

The Predictive Analytics Process

Determine Outcome and Predictors

MeasurementDraw sample,Split into training/holdout

DataData Mining algorithms& Evaluation

Models

Predict New Records;Get More Data;Re-Evaluate

Actions

What to Predict? Why? Implications?

Problem Identification:

Page 8: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

5 Examples of Predictive Analytics

Applications

Page 9: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Problem Identification

Outcome: redemptionPredictors: customer, shop & product info

Measurement

From similar past campaign (redeemers and non-redeemers)

Data

Predictive AlgorithmsExpected gain per offer sent

Models & Evaluation

Example 1:Personalized

Offer

Who to target?

Which coupon?

What medium?

Send Offers (or not!) More Data & Re-Evaluation

Actions

Page 10: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Problem Identification

Outcome: performance Predictors: employee & training info

MeasurementFrom past training efforts (successes and failures)

Data

Which employees to train?

Example 2: Employee Training

Send employees for training (or not!) More Data & Re-Evaluation

Actions

Predictive AlgorithmsExpected gain per employee

Models & Evaluation

Page 11: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Problem Identification

MeasurementOutcome: renewal Predictors: customer & membership info

DataPast renewal campaigns (successes and failures)

Which members most likely not to renew?

Example 3: Customer Churn

Send renewal incentive (or not!) More Data & Re-Evaluation

Actions

Predictive AlgorithmsExpected gain per person

Models & Evaluation

Page 12: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Example 4: Product-level demand forecastingProblem Identification

ActionsUpdate Orders, Pricing, PromoGet More Data, Re-Evaluate

Historic infoData

Forecasting;Expected gain

Models & Eval

MeasurementOutcome: month-ahead weekly forecasts of #units purchased, per itemPredictors: past demand for this & related items, special events, economic outlook, social media

Item-level weekly demand forecasts

Page 13: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Problem Identification

Outcome: pay/not Predictors: customer, product, transaction info

MeasurementPast deliveries (payments and non-payments)

Data

Predict payment probability

Example 5: COD Prediction

Reconfirm with suspect deliveriesMore Data & Update Model

Actions

Predictive AlgorithmsExpected gain per delivery

Models & Evaluation

Page 14: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Predictive Analytics: It’s all about correlation, not causation

Algorithms search for correlation between the outcome and inputs

Different algorithms search for different types of structure – lots of predictive algorithms!

Every time they turn on the seatbelt sign it gets bumpy!

Page 15: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Causality?

www.tylervigen.com

Page 16: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli
Page 17: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

The Causal Explanation Process

Determine Outcome and Causes

MeasurementAssign records to treatment(s)Collect data on inputs+output

DataStatistical models& Evaluation of uncertainty

Models & Eval

Make Decisions; Implement Changes Get More Data and Re-Evaluate

Actions

Which Inputs Cause the Output? How? Implications?Inputs under our control, inputs uncontrollable

Problem Identification:

Page 18: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

What causes average customer to redeem?

Example 1:Personalized Offer

Change coupon design/typeCollect new data (gender)

Actions

Problem Identification:

Tailor trainingPrepare employeesIncentivize learning

Actions

Example 2: Employee Training

What causes average employee to succeed?

Problem Identification:

Page 19: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Improve serviceChange target market

Actions

What causes average member not to renew?

Example 3:Customer Churn

Problem Identification:

Create flexible designsOpen new locations

Actions

Example 4: Demand

Forecasting

What causes high/low demand?

Problem Identification:

Page 20: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Modify payment policyChange website designTrain delivery staff

Actions

What causes average transaction to result in non-payment?

Example 5: Cash-On-Delivery Prediction

Problem Identification:

Page 21: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Toolkit for Determining Causality

Page 22: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Gold Standard: Controlled, Randomized Experiment

Page 23: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Beyond A/B Testing:Multiple factors andInteractions between factors

Page 24: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Causal Explanation withObservational Data

(not a controlled experiment)

Self Selection

Page 25: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Current PracticeCompare online/offline performance stats

Page 26: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Turns out: online and offline users differ on “awareness”

Awareness of electronic services provided by Government of India

Page 27: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Performance Evaluation:% Using Agent

Naïve Comparison:Online system →Less agents

After correcting for self-selection:Online system → More agents for “unaware” users!

Aware Unaware

Page 28: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli
Page 29: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli
Page 30: NASSCOM Big Data and Analytics Summit 2014 - Predicting, Explaining and the Business Analytics Toolkit - Galit Shmuéli

Asia Analytics Lab @ ISBfacebook.com/groups/asiaanalytics