COMM1190-无代写
时间:2023-05-02
COMM1190 Data, Insights and Decisions
Week 8: Research Design &
Experimentation I
General housekeeping for online lecture:
• Please turn on your camera for better engagement
• Please switch your microphone to mute to avoid disruption to the class
• Use your microphone or the chat channel to ask questions or make a comment
• Wait for your lecturer to start
610
Flexibility Week
COMM1190 Data, Insights & Decisions
32 4 5 8 97 101
Intro to
Analytics
Data Exploration
and Visualisation
I
Data Exploration
and Visualisation
II
Predictive
Analytics I
Predictive
Analytics II
Research Design
&
Experimentation
I
Research Design
&
Experimentation
II
Data Ethics Data Communication
x
This week (& next)
 Research design & its importance for prescriptive analytics
 Organisations need to answer what-if & evaluation type questions which involve casual questions
 Experiments using randomised control trials (RCTs) are one way to obtain causal effects
 Big data world associated with many opportunities to run experiments (A/B online experiments)
 Role for non-experimental or observational data
 Data availability provides opportunities to exploit natural experiments (quasi experiments)
Week 8 references
 Accessible if interested
 Fiebig, D.G. (2017), “Big data: Will it improve Patient-Centered Care?”,
The Patient: Patient–Centered Outcomes Research, 10, 133-139.
 Haynes, L., Service, O., Goldacre, B. and Torgenson, B. (2012), “Test, Learn, Adapt: Developing public policy with Randomised Controlled Trials”, Cabinet Office, London.
 Varian, H.R. (2016), “Causal inference in economics and marketing”,
PNAS, 113 (27) 7310-7315.
 Referenced for completeness
 Chattopadhyay, R. and Duflo, E. (2004), “Women as policy makers: Evidence from a randomized policy experiment in India”, Econometrica72, 1409-1443.
Research Design &
Experimentation: Introduction
Motivation
Correlation does not imply causation
 Correlation coefficients measure the strength of an association between two variables
 Recall spurious correlation examples in week 4 – obvious that these do not imply cause & effect
 Other examples may be less obvious to dismiss – there is a correlation between advertising & sales but is it causal?
 How do analysts generate evidence that one action will lead to a particular effect on some outcome of interest?
 How are causal effects reliably estimated?
Introduction
 Organisations require answers to questions & input into decision-making
 Research designStrategy of how one addresses these questions by integrating all parts of the analysis to provide the opportunity to deliver answers
 Constituent parts of data analysis
 Subject matter theory to provide context
 Appropriate data
 Modelling approach that is appropriate for the data & has the
potential to deliver answers – this design question our focus
Introduction…
 What-if & evaluation type questions are crucial for prescriptive analysis
 Involves causal questions requiring estimates of causal effects
 What if a change is made how will that effect future outcomes?
 What impact did an intervention have? Was a policy change that was implemented effective?
 Design particularly important for prescriptive analysis
 Research design: Can causal question being asked be answered by available data & planned modelling approach?
Introduction…
 Case study: Customer churning/retention problem
 Descriptive: Is there a problem with customer churn?
 Predictive: Which customers are at most risk of churning?
 Prescriptive: Which customers are most likely to be
retained if offered incentives to stay? Once implemented
was the incentive cost effective?
 Other examples of such questions
 Q(a) Will it be profitable if on-line advertising is increased?
 Q(b) Will a back-to-work intervention help people get a job?
 Q(c) How much should homeowners living near to a chemical plant be compensated for a chemical spill?
Introduction…
 Experiments are one way to obtain causal effects
 Data deluge in part due to greater opportunities for organizations to collect data & run experiments
 An online A/B experiment could address Q(a)
 Q(b) would require a field experiment
 Some causal questions not amenable to an experiment
 Q(c) is such an example but may be able to get reliable answers from available observational data
 Important design issues related to natural experiments
Introduction…
 Experimental mindset
 Partial equilibrium approach - complex questions broken up into tractable components
 Power of randomisation to control for confounders
 Test, learn, adapt cycle (evidence-based decision-making –as discussed in COMM1110)
Regression: Recall week 4
Regression as an analytical tool
 Linear regression model
= 0 + 1 +
 Useful descriptive device to capture bivariate relationships
 Consider sales () & TV advertising ()(advertising.xlsx)
 Q8.1 What key features of
the data are revealed by
the scatter plot?
0
10
20
30
Sa
le
s
in
th
ou
sa
nd
s
0 100 200 300
Advertising budget in thousands of dollars
sales Fitted values
Sales versus TV advertising
Regression as an analytical tool...
 Have specified a model
 Sales in a market is a linear function of TV advertising
 OLS provides best fit conditional on this model
 Fine as a descriptive device providing stylized facts
 Provides evidence of positive correlation (linear association) between sales & advertising (̂1 = 0.048)
 Prediction of sales for out-of-sample market?
� = 7.033 + 0.048
 A market where = 100 → � = 11.833
Regression as an analytical tool...
 But may want models to do even more – causality & “what-if” counterfactuals
 What happens to sales in a particular market if TV advertising were increased?
 Doesn’t our regression model reliably answer this question?
 At least 2 threats to interpreting ̂1 as causal
 Confounding variables leading to omitted variable bias
 Is estimated advertising effect biased?
 Maybe prices are varying across markets & these are correlated with advertising & hence effects of prices & advertising are confounded
 Reverse causality
 What if markets with low sales increase advertising?
Regression as an analytical tool...
 Prediction models aim to minimize prediction inaccuracy
 Interest is focussed on � not ̂1
 A regression tree could be used provide accurate predictions, but not to address questions of causality
 Questions about causality tend to be more difficult to answer than prediction/forecasting problems
 Need a conceptual basis to guide our approach
 Design issues become very important
Causality & Experimentation
Causality & notion of ceteris paribus
 Causality as defined by philosopher David Lewis:
 “Causation is something that makes a difference, and the
difference it makes must be a difference from what would
have happened without it“
 In evaluating an intervention (or policy change) think of
counterfactual outcomes & what-if questions
 Sales with & without the increase in advertising
 In regression context want to define causal effect of on
 How does variable change if is changed but all other
relevant factors are held constant
 Requires (at a minimum) & to be unrelated
Causality & notion of ceteris paribus...
 Multiple regression provides one approach to better estimate the causal effect of interest (say 1)
= 0 + 11 + 22 + ⋯+ +
 Now more likely 1 & are uncorrelated
 Estimate of 1 controls for other variables & better represents pure impact of 1 (have purged any indirect effect because of correlation with other variables)
 An even better approach is to conduct an experiment
 Randomised control trials (RCTs) suggested as gold standard
 Important to define causal effect of interest & describe how an experiment would be designed to infer causal effect in question
Experiment 1
 Impact of back-to-work program on employment
 “If a person chosen from population of those looking for work is given access to a back-to-work program, will that increase their chance of employment?”
 Implicit assumption: all other factors influencing employment (experience, ability, local employment prospects,...) are held fixed
 Experiment:
 Choose a group of workers looking for work
 Randomly assign them to access the program or not
 Compare employment outcomes in next period
 Experiment works because characteristics of people are unrelated to whether they receive program or not
RCT evaluating back-to-work program
From Haynes et al. (2012)
Experiment 2
 A/B testing of a website landing page
 “If a business rearranges its current website, by how much will this change the conversion rate (new customers)?”
 Implicit assumption: all other factors that influence who visits the website are held fixed
Experiment 2….
 Experiment:
 Design the new webpage
 Randomly assign different users to old (A) & new (B) website
 Compare conversion rates i.e. new customers
 Experiment works because characteristics of users are unrelated to which website is seen
 In online environments relatively easy to conduct
Experiment 3
 Measuring returns to education
 “If a person chosen from the population is given another year of education, by how much will his or her wage increase?”
 Implicit assumption: all other factors that influence wages such as experience, family background, intelligence etc. are held fixed
 Experiment:
 Choose a group of people
 Randomly assign different amounts of education to them!!!
 Compare wage outcomes
 Random assignment is infeasible in this case
 Experiments are not always possible or ethical
Conducting RCTs
 Decide on form of intervention (new program/new website versus status quo)
 Flexibility in designing intervention as new program/website may involve several new features (week 8 workshop, Q1)
 Determine outcome of interest (employment/conversion rates)
 Decide on randomisation unit (workers/customers)
 Determine sample size & randomly assign units to
treatment (new program/new website) & control (no program/old website) groups
 Care required in this step (week 8 workshop, Q2)
Conducting RCTs…
 Compare outcomes to determine treatment effect
 Differences in outcomes can reasonably be attributable to treatment as other aspects of data controlled by researcher
 Decide on whether to adapt (implement program/use new website) or not on basis of findings
 See asynchronous lecture (Experiments in a Big Data World) for Case Study 2 & week 8 workshop, Q3
Experimental evidence
 Experimental evidence is input into decision-making
 Even if RCT yields a significant treatment effect this may not be enough to justify implimentation of intervention
 Could it have harmful unintended consequences?
 Does intervention represent value for money?
 Interventions are costly so size of any benefit needs to be weighed against the cost
 Null results are useful!
 A RCT that does not provide evidence of a treatment effect could avoid uneccesary costs
Experimental evidence…
 Interventions may be intuitively appealing
 But typically many intuitively appealing interventions
 Which is better & whether they are value for money requires supporting evidence
 Once implemented evidence should continue to be collected & interventions refined where appropriate
 Interventions may work in one population but not another
 Replication of core findings across experiments represents more compelling evidence than a very significant effect in a single study
Observational data
Observational data versus
experimental data
 Why experiment when observational data are abundant?
 “…it is invariably true that the
available data have not been
collected with research in mind.
… there is a mismatch between
key concepts and available data
or what is available suffers from
sample selection problems. ” Fiebig(2017)
Observational data versus experimental
data…
 In observational data outcomes represent actual behaviour
 Employment prospects depend on personal characteristics, employment conditions, ...
 Web access depends on personal characteristics, other advertising campaigns, ...
 No reason why factors impacting outcome are randomly assigned
 Did people who were given the back-to-work program already have better employment propects?
 This is problem of confounding (lurking) variables
 “The economy is a miserable experimental design”
Robert E. Lucas Jr.
Observational data versus experimental data…
Observational data versus experimental
data…
 What if workers choose to participate in the program?
 If workers base their choice on likely benefits from program then likely to provide an inflated (biased) estimate of the treatment effect
 This is selection bias induced by an endogenous treatment
 Q8.2 Does SAS example with huge sample & many controls avoid
this selection problem?
 Random assignment avoids this selection problem
 Bottom line, useful to think with an experimental mindset
Potential outcomes framework
Recall linear regression model of week 4
 Simple linear regression model
= 0 + 1 +
 Useful special case if represents a binary treatment
 Worker receives program ( = 1) or not ( = 0)
 Customer sees new website ( = 1) or old ( = 0)
 Patient received drug ( = 1) or placebo ( = 0)
 More generally simply denotes group membership
 = 1 if female, = 0 otherwise
 Here gender is not a treatment!
Interpretation of estimates?
 When simple linear regression contains a binary regressor
= 0 + 1 +
 Then straightforward to show that OLS estimates are
̂0 = �(0) & ̂1 = �(1) − �(0)
where � is mean of conditional on
̂1 is difference in means (difference in proportions when is binary)
Potential outcomes
 Simple regression of outcomes on binary treatment recovers difference in means between treated & controls
 Difference in proportions in Experiments 1 & 2
 Difference is descriptive but can it be interpreted as causal?
 Potential outcomes framework considers outcomes in two states of the world (1 if treated or 0 if not)
 Treatment effect is difference in two potential outcomes
= 1 − 0
 Q8.3 Without more structure the treatment effect can’t be
estimated. Why?
Potential outcomes....
 Fundamental problem of causal inference – need a credible way to infer unobserved counterfactual outcomes
 If focus on average treatment effect (ATE)
= () = 1 − 0
 … and assume random treatment assignment
 … then in = 0 + 1 + , 1 = ATE
 Implying estimated ATE is the difference-in-means estimator
Case study#1:
Women as policy makers
Case study#1: Background
 Data collection costs have declined dramatically
 Field experiments undertaking RCTs now more prevalent in business & economics
 2019 Nobel Prize in Economics for “…their new experiment-based approach has transformed development economics”
 Chattopadhyay & Duflo (2004) estimated casual effect of having female politicians in government on policy outcomes
 For a period in India a third of village council heads were randomly reserved for female politicians
Case study#1: Background…
 Data women.csv includes 322 village-level observations
 Hypothesis is that women leaders will support policies that women voters care about more
 reserved indicates reserved for a female village leader
 female indicates a female village leader
 irrigation & water are number of new or repaired facilities of this type – men tended to be more concerned about former & women the later
 Q8.4 Why would an observational study be problematic in
determining whether women politicians promote
different policies when in government?
Case study#1: Estimates
 First it seems that the policy was successfully applied
 All 108 treated ( = 1) villages had a female head
 Only 16 of the 224 control villages had a female head
 Estimate regression models
 Confirm in week 8 workshop, Q4
= 0 + 1 +
 Estimated ATE is 9.25 increase in projects attributable to the policy & precisely estimated with 95% CI [1.49 , 17.02]
= α0 + α1 +
 Estimated ATE is −0.37 decrease in projects attributable to the policy but imprecise with 95% CI [−2.58 , 1.84]
Case study#1: Bottom line
 ATEs indicate support for the hypothesis
 Treated villages with female heads were much more likely to support water projects
 Treated villages with female heads were less likely to support irrigation projects but this effect had a 95% CI overlapping zero
Progress report #1
 Data deluge has extended opportunities for data analysis & questions that can be answered
 Stressed role of research design in combination with available data
 Stressed role of regression as primary tool of analysis
 Previously used in description & prediction
 Have been more explicit about what constitutes evidence of causation & threats posed by confoundment & selection
 Important role for experiments
 In week 9 expand on role of observational data
 Using such data comes with threats that need to be recognized
Thank you
Remember to watch asynchronous
lecture before workshop this week
Direct questions on lecture material to:
Professor Denzil Fiebig
d.fiebig@unsw.edu.au
Consultation:
BUS444 Mon 3-4:30, Tues 10-11:30
essay、essay代写