SlideShare una empresa de Scribd logo
1 de 33
Descargar para leer sin conexión
Social foundations for statistics and machine learning
Maximilian Kasy
September 23, 2021
Today’s argument
Single-agent decision theory provides the foundation
for both statistics and machine learning (ML).
• Examples of decision problems:
• Experimental design, estimation, inference, policy choice,
• supervised learning, adaptive treatment choice.
• Components of decision problems:
• Unknown state of the world,
• which impacts the distribution of observed data,
• as well as the loss function used to ultimately evaluate the decision.
1 / 28
The limits of decision theory
• This single-agent framework provides important insights.
• But it cannot address important scientific and societal challenges.
1. Replication crisis, publication bias, p-hacking, pre-registration,
reforms of statistics teaching and the publication system.
2. The social impact of AI,
algorithmic discrimination and inequality,
value alignment of autonomous agents / robots.
2 / 28
Science and technology are social activities
• Scientific knowledge production, and the deployment of technology,
are inherently social:
1. Different agents have different objectives (“loss functions”).
2. The objectives of statistics and ML
are socially determined – who’s objectives matter?
3. Replication crisis and reform proposals, and conflicts over the impact of AI
can only be understood if we take this into account.
• This is well recognized by the humanities:
• Philosophy, sociology, and history of science, science and technology studies.
• But these fields do not provide formal and prescriptive recommendations
for quantitative empirical researchers, or AI engineers.
3 / 28
Economics to the rescue
• Economics is well positioned to fill this gap:
• We share the languages of constrained optimization and probability theory
with statistics and ML.
• But we are also used to considering multiple agents
with unequal endowments, conflicting interests, private information.
• Today, I discuss two projects that are part of this general agenda:
Kasy, M. and Spiess, J. (2021). Rationalizing pre-analysis plans: Statisti-
cal decisions subject to implementability.
(work in progress)
Kasy, M. and Abebe, R. (2021). Fairness, equality, and power in algorith-
mic decision making.
(published, FAccT 2021)
4 / 28
Decision theory – a quick review
P-hacking and pre-analysis plans
Algorithmic fairness and economic inequality
Conclusion
AI as decision theory
The textbook “Artificial intelligence: a modern approach” (Russell and Norvig, 2016)
defines the goal of AI as the derivation of
“ general principles of rational agents and on components for constructing
them.”
“An agent is anything that can be viewed as perceiving its environment
through sensors and acting upon that environment through actuators.”
“For each possible percept sequence, a rational agent should select an ac-
tion that is expected to maximize its performance measure, given
the evidence provided by the percept sequence and whatever built-in
knowledge the agent has.”
5 / 28
Statistics as decision theory
Similarly, the Bayesian statistics textbook by Robert (2007) states:
“ Considering that the overall purpose of most inferential studies is to pro-
vide the statistician (or a client) with a decision, it seems reasonable to
ask for an evaluation criterion of decision procedures that assesses
the consequences of each decision and depends on the parameters
of the model, i.e., the true state of the world (or of Nature).”
“ [...] implies a reinforced axiomatization of the statistical inferential
framework, called Decision Theory. This augmented theoretical struc-
ture is necessary for Statistics to reach a coherence otherwise unattain-
able.”
6 / 28
Decision theory – General setup
state of the world
θ
observed data
X
decision
a
loss
L(a,θ)
decision function
a=δ(X)
statistical
model
X~f(x,θ)
7 / 28
Examples of decision problems
• Estimation:
• Find an a which is close to some function µ of θ.
• Typical loss function: L(a, θ) = (a − µ(θ))2
.
• Testing:
• Decide whether H0 : θ ∈ Θ0 is true.
• Typical loss function: L(a, θ) = 1(a = 1, θ ∈ Θ0) + c · 1(a = 0, θ /
∈ Θ0).
• Targeted treatment assignment:
• Assign treatment W as a function of features X, W = δ(X).
• Typical utility function: E[δ(X) · (M − c)],
for treatment effects M, treatment cost c.
8 / 28
Decision theory – a quick review
P-hacking and pre-analysis plans
Algorithmic fairness and economic inequality
Conclusion
P-hacking and pre-analysis plans
• Trial registration and pre-analysis plans (PAPs):
A standard requirement for experimental research.
• Clinical studies / medicine: Starting in the 1990s.
• (Field) experiments in economics: More recently.
• Standard justification: Guarantee validity of inference.
• P-hacking, specification searching, and selective publication distort inference.
• Tying researchers’ hands prevents selective reporting.
• “PAPs are to frequentist inference what RCTs are to causality.”
• Counter-arguments:
• Pre-specification is costly.
• Interesting findings are unexpected and flexibility is necessary.
9 / 28
No commitment (pre-registration) in decision theory
• Two alternatives, in a generic decision problem:
1. We can commit to (pre-register) a rule δ(·) before observing X.
2. We can pick δ(X) after observing X.
• By the law of iterated expectations
E[L(δ(X), θ)] = E[E[L(δ(X), θ)|X]]
=
X
x
E[L(δ(x), θ)|X = x] · P(X = x).
• Therefore:
• Picking the optimal δ(·) (to minimize the sum) is the same
• as picking the optimal δ(x) for every value of x (each term of the sum).
• The decision-problem is dynamically consistent.
10 / 28
A mechanism design perspective
Claim: Concerns about p-hacking, publication bias, pre-registration
are at their core about divergent interests between multiple actors.
Q: How to incorporate this social dimension into prescriptive methodology?
A: Model statistical inference as a mechanism design problem!
• Take the perspective of a reader of empirical research
who wants to implement a statistical decision rule.
• Not all rules are implementable when researchers have
divergent interests and private information about the data,
and they can selectively report to readers.
• Agenda: Characterize optimal decision rules subject to implementability.
11 / 28
Setup
• Two agents: Decision-maker and analyst.
• The analyst observes a vector
X = (X1, . . . , Xn̄),
where
Xi
iid
∼ Ber(θ).
• Analyst: Reports a subvector XI to the decision-maker, where
I ⊂ {1, . . . , n̄}.
• Decision-maker: Makes a decision
a ∈ {0, 1},
based on this report.
12 / 28
Prior and objectives
• Common prior:
θ ∼ Beta(α, β).
• Analyst’s objective:
uan
= a − c · |I|.
|I| is the size of the reported set,
c is the cost of communicating an additional component.
• Decision-maker’s objective:
ud-m
= a · (θ − θ).
θ is a commonly known parameter.
Minimum value of θ beyond which the decision-maker would like to choose
a = 1.
13 / 28
Timeline
1. The decision-maker commits to a decision rule
a = a(J, I, XI).
2. The analyst reports a PAP
J ⊆ {1, . . . , n̄}.
3. The analyst next observes X, chooses I ⊆ {1, . . . , n̄}, and reports
(I, XI).
4. The decision rule is applied and utilities are realized.
14 / 28
Implementability
• Let x denote values that the random vector X may take.
• Reduced form mapping (statistical decision rule)
x 7→ ā(x).
• ā(x) is implementable
if there exist mappings I(x) and a(I, xI)
such that for all x
ā(x) = a(I(x), xI(x)),
and
I(x) ∈ argmax
I
a(I, xI) − c · |I|.
15 / 28
Notation
• Successes among all components: s(X) =
Pn̄
i=1 Xi.
Successes among the subset I: s(XI) =
P
i∈I Xi.
Failures among the subset I: t(XI) = |I| − s(XI).
• Maximal number of components the analyst is willing to submit:
n̄PC
= max {n : 1 − cn ≥ 0} = b1/cc .
• First-best cutoff for the decision-maker:
s∗
(n) = min

s : E[θ
Social foundations slides (1)
s(X1,...,n)=s] ≥ θ
	
.
• Minimal cutoff for the decision-maker:
smin
(n) = min

s : E[θ
Social foundations slides (1)
s(X1,...,n)≥s] ≥ θ
	
.
16 / 28
Symmetric decision rules
• Consider symmetric rules of the form
a(s(XI), t(XI)),
Proposition (Optimal symmetric decision rule)
The optimal reduced-form decision rule that is symmetrically implementable
takes the form
ā = 1(s(X) ≥ min(s∗
, n̄PC
)),
if n̄PC ≥ smin, and can be implemented by
a = 1(s(XI) ≥ min(s∗
, n̄PC
)).
Otherwise the optimal decision rule is given by a ≡ 0.
17 / 28
Symmetric cutoff without PAP, uniform prior
If the number of components n̄ is to the right of the maximum n̄∗:
• PAPs increase decision-maker welfare
• by forcing the analyst to ignore all components i  n̄∗.
18 / 28
Decision theory – a quick review
P-hacking and pre-analysis plans
Algorithmic fairness and economic inequality
Conclusion
Algorithmic fairness and economic inequality
1. Standard definitions of algorithmic fairness:
• Absence of “bias.”
• Similar to “taste based discrimination” in economics,
which is defined as a deviation from profit maximization.
• Fairness as a decision problem, aligning treatment and latent merit.
2. Social choice theory (in economics), theory of justice (in political philosophy):
• Social welfare is typically defined based on individuals’ welfare.
• Points of contention:
How to measure individual welfare,
how to aggregate / trade off welfare across individuals.
• Policies (and algorithms!) are evaluated
based on their consequences for social welfare.
“Bias” versus “social welfare” have very different implications!
19 / 28
Fairness in algorithmic decision making – Setup
• Binary treatment W, treatment return M (heterogeneous), treatment cost c.
Decision maker’s objective
µ = E[W · (M − c)].
• All expectations denote averages across individuals (not uncertainty).
• M is unobserved, but predictable based on features X.
For m(x) = E[M|X = x], the optimal policy is
w∗
(x) = 1(m(X)  c).
20 / 28
Examples
• Bail setting for defendants based on predicted recidivism.
• Screening of job candidates based on predicted performance.
• Consumer credit based on predicted repayment.
• Screening of tenants for housing based on predicted payment risk.
• Admission to schools based on standardized tests.
21 / 28
Definitions of fairness
• Most definitions depend on three ingredients.
1. Treatment W (job, credit, incarceration, school admission).
2. A notion of merit M (marginal product, credit default, recidivism, test
performance).
3. Protected categories A (ethnicity, gender).
• I will focus on the following definition of fairness:
π = E[M|W = 1, A = 1] − E[M|W = 1, A = 0] = 0
“Average merit, among the treated, does not vary across the groups a.”
This is called “predictive parity” in ML,
the “hit rate test” for “taste based discrimination” in economics.
22 / 28
Fairness and D’s objective
Observation
Suppose that W, M are binary (“classification”), and that
1. m(X) = M (perfect predictability), and
2. w∗(x) = 1(m(X)  c) (unconstrained maximization of D’s objective µ).
Then w∗(x) satisfies predictive parity, i.e., π = 0.
In words:
• If D is a firm that is maximizing profits and observes everything
then their decisions are fair by assumption.
– No matter how unequal the resulting outcomes within and across groups.
• Only deviations from profit-maximization are “unfair.”
23 / 28
Three normative limitations of “fairness” as predictive parity
Notions of fairness of this form have several key limitations:
1. They legitimize and perpetuate inequalities justified by “merit.”
Where does inequality in M come from?
2. They are narrowly bracketed.
Inequality in W in the algorithm,
instead of some outcomes Y in a wider population.
3. Fairness-based perspectives focus on categories (protected groups)
and ignore within-group inequality.
24 / 28
Social welfare as an alternative framework
• The framework of fairness / bias / discrimination contrasts with
perspectives focused on consequences for social welfare.
• Common presumption for most theories of justice:
Normative statements about society
are based on statements about individual welfare
• Formally:
• Individuals i = 1, . . . , n
• Individual i’s welfare Yi
• Social welfare as function of individuals’ welfare
SWF = F(Y1, . . . , Yn).
• Key points of contention:
1. Who is included among the individuals i? Who’s lives matter?
2. How to measure individual welfare Yi?
3. How to trade off welfare across individuals i?
25 / 28

Más contenido relacionado

Más de Office for National Statistics

So what does ‘Beyond GDP’ mean for the UK – 12 October 2023
So what does ‘Beyond GDP’ mean for the UK – 12 October 2023So what does ‘Beyond GDP’ mean for the UK – 12 October 2023
So what does ‘Beyond GDP’ mean for the UK – 12 October 2023Office for National Statistics
 
GDP after 2025: updating national accounts and balance of payments – 11 Octob...
GDP after 2025: updating national accounts and balance of payments – 11 Octob...GDP after 2025: updating national accounts and balance of payments – 11 Octob...
GDP after 2025: updating national accounts and balance of payments – 11 Octob...Office for National Statistics
 
SlideShare Measuring the Economy Slidepack - 29 September 2023
SlideShare Measuring the Economy Slidepack - 29 September 2023SlideShare Measuring the Economy Slidepack - 29 September 2023
SlideShare Measuring the Economy Slidepack - 29 September 2023Office for National Statistics
 
SlideShare ONS Economic Forum Slidepack - 18 September 2023
SlideShare ONS Economic Forum Slidepack - 18 September 2023SlideShare ONS Economic Forum Slidepack - 18 September 2023
SlideShare ONS Economic Forum Slidepack - 18 September 2023Office for National Statistics
 
ONS Local presents Suffolk County Council's Cost of Living Dashboard
ONS Local presents Suffolk County Council's Cost of Living DashboardONS Local presents Suffolk County Council's Cost of Living Dashboard
ONS Local presents Suffolk County Council's Cost of Living DashboardOffice for National Statistics
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsOffice for National Statistics
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsOffice for National Statistics
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsOffice for National Statistics
 
ONS Local presents: Adult Education Outcomes in London
ONS Local presents: Adult Education Outcomes in LondonONS Local presents: Adult Education Outcomes in London
ONS Local presents: Adult Education Outcomes in LondonOffice for National Statistics
 
SlideShare ONS Economic Forum Slidepack - 12 June 2023
SlideShare ONS Economic Forum Slidepack - 12 June 2023SlideShare ONS Economic Forum Slidepack - 12 June 2023
SlideShare ONS Economic Forum Slidepack - 12 June 2023Office for National Statistics
 
ONS Local presents: Census 2021, products and analysis
ONS Local presents: Census 2021, products and analysisONS Local presents: Census 2021, products and analysis
ONS Local presents: Census 2021, products and analysisOffice for National Statistics
 
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptx
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptxSlideShare ONS Economic Forum Slidepack - 15 May 2023.pptx
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptxOffice for National Statistics
 
SlideShare Annual crime and justice statistics forum 2023.pptx
SlideShare Annual crime and justice statistics forum 2023.pptxSlideShare Annual crime and justice statistics forum 2023.pptx
SlideShare Annual crime and justice statistics forum 2023.pptxOffice for National Statistics
 

Más de Office for National Statistics (20)

So what does ‘Beyond GDP’ mean for the UK – 12 October 2023
So what does ‘Beyond GDP’ mean for the UK – 12 October 2023So what does ‘Beyond GDP’ mean for the UK – 12 October 2023
So what does ‘Beyond GDP’ mean for the UK – 12 October 2023
 
GDP after 2025: updating national accounts and balance of payments – 11 Octob...
GDP after 2025: updating national accounts and balance of payments – 11 Octob...GDP after 2025: updating national accounts and balance of payments – 11 Octob...
GDP after 2025: updating national accounts and balance of payments – 11 Octob...
 
SlideShare Measuring the Economy Slidepack - 29 September 2023
SlideShare Measuring the Economy Slidepack - 29 September 2023SlideShare Measuring the Economy Slidepack - 29 September 2023
SlideShare Measuring the Economy Slidepack - 29 September 2023
 
Why dashboards?
Why dashboards?Why dashboards?
Why dashboards?
 
SlideShare ONS Economic Forum Slidepack - 18 September 2023
SlideShare ONS Economic Forum Slidepack - 18 September 2023SlideShare ONS Economic Forum Slidepack - 18 September 2023
SlideShare ONS Economic Forum Slidepack - 18 September 2023
 
Connecting to the StatXplore API in PowerBI
Connecting to the StatXplore API in PowerBIConnecting to the StatXplore API in PowerBI
Connecting to the StatXplore API in PowerBI
 
ONS Local presents Suffolk County Council's Cost of Living Dashboard
ONS Local presents Suffolk County Council's Cost of Living DashboardONS Local presents Suffolk County Council's Cost of Living Dashboard
ONS Local presents Suffolk County Council's Cost of Living Dashboard
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIs
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIs
 
ONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIsONS Local and Data Science Community Workshop 1: How to use APIs
ONS Local and Data Science Community Workshop 1: How to use APIs
 
ONS Local presents: Adult Education Outcomes in London
ONS Local presents: Adult Education Outcomes in LondonONS Local presents: Adult Education Outcomes in London
ONS Local presents: Adult Education Outcomes in London
 
ONS Local presents: Explore Subnational Statistics
ONS Local presents: Explore Subnational StatisticsONS Local presents: Explore Subnational Statistics
ONS Local presents: Explore Subnational Statistics
 
ONS Local presents - Census 2021 Education Analysis
ONS Local presents - Census 2021 Education AnalysisONS Local presents - Census 2021 Education Analysis
ONS Local presents - Census 2021 Education Analysis
 
SlideShare ONS Economic Forum Slidepack - 12 June 2023
SlideShare ONS Economic Forum Slidepack - 12 June 2023SlideShare ONS Economic Forum Slidepack - 12 June 2023
SlideShare ONS Economic Forum Slidepack - 12 June 2023
 
ONS local presents clustering
ONS local presents clusteringONS local presents clustering
ONS local presents clustering
 
ONS Local presents: Census 2021, products and analysis
ONS Local presents: Census 2021, products and analysisONS Local presents: Census 2021, products and analysis
ONS Local presents: Census 2021, products and analysis
 
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptx
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptxSlideShare ONS Economic Forum Slidepack - 15 May 2023.pptx
SlideShare ONS Economic Forum Slidepack - 15 May 2023.pptx
 
SlideShare Annual crime and justice statistics forum 2023.pptx
SlideShare Annual crime and justice statistics forum 2023.pptxSlideShare Annual crime and justice statistics forum 2023.pptx
SlideShare Annual crime and justice statistics forum 2023.pptx
 
ONS business data and statistics user event
ONS business data and statistics user eventONS business data and statistics user event
ONS business data and statistics user event
 
ONS Economic Forum, 17 April 2023.pptx
ONS Economic Forum, 17 April 2023.pptxONS Economic Forum, 17 April 2023.pptx
ONS Economic Forum, 17 April 2023.pptx
 

Último

Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023ymrp368
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxolyaivanovalion
 
Midocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxMidocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxolyaivanovalion
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxolyaivanovalion
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxJohnnyPlasten
 
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779Best VIP Call Girls Noida Sector 39 Call Me: 8448380779
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779Delhi Call girls
 
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxBPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxMohammedJunaid861692
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxolyaivanovalion
 
Zuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxZuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxolyaivanovalion
 
Discover Why Less is More in B2B Research
Discover Why Less is More in B2B ResearchDiscover Why Less is More in B2B Research
Discover Why Less is More in B2B Researchmichael115558
 
Smarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxSmarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxolyaivanovalion
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxolyaivanovalion
 
Determinants of health, dimensions of health, positive health and spectrum of...
Determinants of health, dimensions of health, positive health and spectrum of...Determinants of health, dimensions of health, positive health and spectrum of...
Determinants of health, dimensions of health, positive health and spectrum of...shambhavirathore45
 
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...Delhi Call girls
 
Invezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz1
 
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Delhi Call girls
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% SecurePooja Nehwal
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionfulawalesam
 

Último (20)

Data-Analysis for Chicago Crime Data 2023
Data-Analysis for Chicago Crime Data  2023Data-Analysis for Chicago Crime Data  2023
Data-Analysis for Chicago Crime Data 2023
 
BabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptxBabyOno dropshipping via API with DroFx.pptx
BabyOno dropshipping via API with DroFx.pptx
 
Midocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFxMidocean dropshipping via API with DroFx
Midocean dropshipping via API with DroFx
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptx
 
Log Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptxLog Analysis using OSSEC sasoasasasas.pptx
Log Analysis using OSSEC sasoasasasas.pptx
 
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779Best VIP Call Girls Noida Sector 39 Call Me: 8448380779
Best VIP Call Girls Noida Sector 39 Call Me: 8448380779
 
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxBPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
 
Mature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptxMature dropshipping via API with DroFx.pptx
Mature dropshipping via API with DroFx.pptx
 
Zuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptxZuja dropshipping via API with DroFx.pptx
Zuja dropshipping via API with DroFx.pptx
 
Discover Why Less is More in B2B Research
Discover Why Less is More in B2B ResearchDiscover Why Less is More in B2B Research
Discover Why Less is More in B2B Research
 
Smarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxSmarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptx
 
CebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptxCebaBaby dropshipping via API with DroFX.pptx
CebaBaby dropshipping via API with DroFX.pptx
 
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICECHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
CHEAP Call Girls in Saket (-DELHI )🔝 9953056974🔝(=)/CALL GIRLS SERVICE
 
Determinants of health, dimensions of health, positive health and spectrum of...
Determinants of health, dimensions of health, positive health and spectrum of...Determinants of health, dimensions of health, positive health and spectrum of...
Determinants of health, dimensions of health, positive health and spectrum of...
 
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
Call Girls in Sarai Kale Khan Delhi 💯 Call Us 🔝9205541914 🔝( Delhi) Escorts S...
 
Invezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signalsInvezz.com - Grow your wealth with trading signals
Invezz.com - Grow your wealth with trading signals
 
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
Best VIP Call Girls Noida Sector 22 Call Me: 8448380779
 
Sampling (random) method and Non random.ppt
Sampling (random) method and Non random.pptSampling (random) method and Non random.ppt
Sampling (random) method and Non random.ppt
 
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% SecureCall me @ 9892124323  Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
Call me @ 9892124323 Cheap Rate Call Girls in Vashi with Real Photo 100% Secure
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interaction
 

Social foundations slides (1)

  • 1. Social foundations for statistics and machine learning Maximilian Kasy September 23, 2021
  • 2. Today’s argument Single-agent decision theory provides the foundation for both statistics and machine learning (ML). • Examples of decision problems: • Experimental design, estimation, inference, policy choice, • supervised learning, adaptive treatment choice. • Components of decision problems: • Unknown state of the world, • which impacts the distribution of observed data, • as well as the loss function used to ultimately evaluate the decision. 1 / 28
  • 3. The limits of decision theory • This single-agent framework provides important insights. • But it cannot address important scientific and societal challenges. 1. Replication crisis, publication bias, p-hacking, pre-registration, reforms of statistics teaching and the publication system. 2. The social impact of AI, algorithmic discrimination and inequality, value alignment of autonomous agents / robots. 2 / 28
  • 4. Science and technology are social activities • Scientific knowledge production, and the deployment of technology, are inherently social: 1. Different agents have different objectives (“loss functions”). 2. The objectives of statistics and ML are socially determined – who’s objectives matter? 3. Replication crisis and reform proposals, and conflicts over the impact of AI can only be understood if we take this into account. • This is well recognized by the humanities: • Philosophy, sociology, and history of science, science and technology studies. • But these fields do not provide formal and prescriptive recommendations for quantitative empirical researchers, or AI engineers. 3 / 28
  • 5. Economics to the rescue • Economics is well positioned to fill this gap: • We share the languages of constrained optimization and probability theory with statistics and ML. • But we are also used to considering multiple agents with unequal endowments, conflicting interests, private information. • Today, I discuss two projects that are part of this general agenda: Kasy, M. and Spiess, J. (2021). Rationalizing pre-analysis plans: Statisti- cal decisions subject to implementability. (work in progress) Kasy, M. and Abebe, R. (2021). Fairness, equality, and power in algorith- mic decision making. (published, FAccT 2021) 4 / 28
  • 6. Decision theory – a quick review P-hacking and pre-analysis plans Algorithmic fairness and economic inequality Conclusion
  • 7. AI as decision theory The textbook “Artificial intelligence: a modern approach” (Russell and Norvig, 2016) defines the goal of AI as the derivation of “ general principles of rational agents and on components for constructing them.” “An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators.” “For each possible percept sequence, a rational agent should select an ac- tion that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has.” 5 / 28
  • 8. Statistics as decision theory Similarly, the Bayesian statistics textbook by Robert (2007) states: “ Considering that the overall purpose of most inferential studies is to pro- vide the statistician (or a client) with a decision, it seems reasonable to ask for an evaluation criterion of decision procedures that assesses the consequences of each decision and depends on the parameters of the model, i.e., the true state of the world (or of Nature).” “ [...] implies a reinforced axiomatization of the statistical inferential framework, called Decision Theory. This augmented theoretical struc- ture is necessary for Statistics to reach a coherence otherwise unattain- able.” 6 / 28
  • 9. Decision theory – General setup state of the world θ observed data X decision a loss L(a,θ) decision function a=δ(X) statistical model X~f(x,θ) 7 / 28
  • 10. Examples of decision problems • Estimation: • Find an a which is close to some function µ of θ. • Typical loss function: L(a, θ) = (a − µ(θ))2 . • Testing: • Decide whether H0 : θ ∈ Θ0 is true. • Typical loss function: L(a, θ) = 1(a = 1, θ ∈ Θ0) + c · 1(a = 0, θ / ∈ Θ0). • Targeted treatment assignment: • Assign treatment W as a function of features X, W = δ(X). • Typical utility function: E[δ(X) · (M − c)], for treatment effects M, treatment cost c. 8 / 28
  • 11. Decision theory – a quick review P-hacking and pre-analysis plans Algorithmic fairness and economic inequality Conclusion
  • 12. P-hacking and pre-analysis plans • Trial registration and pre-analysis plans (PAPs): A standard requirement for experimental research. • Clinical studies / medicine: Starting in the 1990s. • (Field) experiments in economics: More recently. • Standard justification: Guarantee validity of inference. • P-hacking, specification searching, and selective publication distort inference. • Tying researchers’ hands prevents selective reporting. • “PAPs are to frequentist inference what RCTs are to causality.” • Counter-arguments: • Pre-specification is costly. • Interesting findings are unexpected and flexibility is necessary. 9 / 28
  • 13. No commitment (pre-registration) in decision theory • Two alternatives, in a generic decision problem: 1. We can commit to (pre-register) a rule δ(·) before observing X. 2. We can pick δ(X) after observing X. • By the law of iterated expectations E[L(δ(X), θ)] = E[E[L(δ(X), θ)|X]] = X x E[L(δ(x), θ)|X = x] · P(X = x). • Therefore: • Picking the optimal δ(·) (to minimize the sum) is the same • as picking the optimal δ(x) for every value of x (each term of the sum). • The decision-problem is dynamically consistent. 10 / 28
  • 14. A mechanism design perspective Claim: Concerns about p-hacking, publication bias, pre-registration are at their core about divergent interests between multiple actors. Q: How to incorporate this social dimension into prescriptive methodology? A: Model statistical inference as a mechanism design problem! • Take the perspective of a reader of empirical research who wants to implement a statistical decision rule. • Not all rules are implementable when researchers have divergent interests and private information about the data, and they can selectively report to readers. • Agenda: Characterize optimal decision rules subject to implementability. 11 / 28
  • 15. Setup • Two agents: Decision-maker and analyst. • The analyst observes a vector X = (X1, . . . , Xn̄), where Xi iid ∼ Ber(θ). • Analyst: Reports a subvector XI to the decision-maker, where I ⊂ {1, . . . , n̄}. • Decision-maker: Makes a decision a ∈ {0, 1}, based on this report. 12 / 28
  • 16. Prior and objectives • Common prior: θ ∼ Beta(α, β). • Analyst’s objective: uan = a − c · |I|. |I| is the size of the reported set, c is the cost of communicating an additional component. • Decision-maker’s objective: ud-m = a · (θ − θ). θ is a commonly known parameter. Minimum value of θ beyond which the decision-maker would like to choose a = 1. 13 / 28
  • 17. Timeline 1. The decision-maker commits to a decision rule a = a(J, I, XI). 2. The analyst reports a PAP J ⊆ {1, . . . , n̄}. 3. The analyst next observes X, chooses I ⊆ {1, . . . , n̄}, and reports (I, XI). 4. The decision rule is applied and utilities are realized. 14 / 28
  • 18. Implementability • Let x denote values that the random vector X may take. • Reduced form mapping (statistical decision rule) x 7→ ā(x). • ā(x) is implementable if there exist mappings I(x) and a(I, xI) such that for all x ā(x) = a(I(x), xI(x)), and I(x) ∈ argmax I a(I, xI) − c · |I|. 15 / 28
  • 19. Notation • Successes among all components: s(X) = Pn̄ i=1 Xi. Successes among the subset I: s(XI) = P i∈I Xi. Failures among the subset I: t(XI) = |I| − s(XI). • Maximal number of components the analyst is willing to submit: n̄PC = max {n : 1 − cn ≥ 0} = b1/cc . • First-best cutoff for the decision-maker: s∗ (n) = min s : E[θ
  • 21. s(X1,...,n)=s] ≥ θ . • Minimal cutoff for the decision-maker: smin (n) = min s : E[θ
  • 24. Symmetric decision rules • Consider symmetric rules of the form a(s(XI), t(XI)), Proposition (Optimal symmetric decision rule) The optimal reduced-form decision rule that is symmetrically implementable takes the form ā = 1(s(X) ≥ min(s∗ , n̄PC )), if n̄PC ≥ smin, and can be implemented by a = 1(s(XI) ≥ min(s∗ , n̄PC )). Otherwise the optimal decision rule is given by a ≡ 0. 17 / 28
  • 25. Symmetric cutoff without PAP, uniform prior If the number of components n̄ is to the right of the maximum n̄∗: • PAPs increase decision-maker welfare • by forcing the analyst to ignore all components i n̄∗. 18 / 28
  • 26. Decision theory – a quick review P-hacking and pre-analysis plans Algorithmic fairness and economic inequality Conclusion
  • 27. Algorithmic fairness and economic inequality 1. Standard definitions of algorithmic fairness: • Absence of “bias.” • Similar to “taste based discrimination” in economics, which is defined as a deviation from profit maximization. • Fairness as a decision problem, aligning treatment and latent merit. 2. Social choice theory (in economics), theory of justice (in political philosophy): • Social welfare is typically defined based on individuals’ welfare. • Points of contention: How to measure individual welfare, how to aggregate / trade off welfare across individuals. • Policies (and algorithms!) are evaluated based on their consequences for social welfare. “Bias” versus “social welfare” have very different implications! 19 / 28
  • 28. Fairness in algorithmic decision making – Setup • Binary treatment W, treatment return M (heterogeneous), treatment cost c. Decision maker’s objective µ = E[W · (M − c)]. • All expectations denote averages across individuals (not uncertainty). • M is unobserved, but predictable based on features X. For m(x) = E[M|X = x], the optimal policy is w∗ (x) = 1(m(X) c). 20 / 28
  • 29. Examples • Bail setting for defendants based on predicted recidivism. • Screening of job candidates based on predicted performance. • Consumer credit based on predicted repayment. • Screening of tenants for housing based on predicted payment risk. • Admission to schools based on standardized tests. 21 / 28
  • 30. Definitions of fairness • Most definitions depend on three ingredients. 1. Treatment W (job, credit, incarceration, school admission). 2. A notion of merit M (marginal product, credit default, recidivism, test performance). 3. Protected categories A (ethnicity, gender). • I will focus on the following definition of fairness: π = E[M|W = 1, A = 1] − E[M|W = 1, A = 0] = 0 “Average merit, among the treated, does not vary across the groups a.” This is called “predictive parity” in ML, the “hit rate test” for “taste based discrimination” in economics. 22 / 28
  • 31. Fairness and D’s objective Observation Suppose that W, M are binary (“classification”), and that 1. m(X) = M (perfect predictability), and 2. w∗(x) = 1(m(X) c) (unconstrained maximization of D’s objective µ). Then w∗(x) satisfies predictive parity, i.e., π = 0. In words: • If D is a firm that is maximizing profits and observes everything then their decisions are fair by assumption. – No matter how unequal the resulting outcomes within and across groups. • Only deviations from profit-maximization are “unfair.” 23 / 28
  • 32. Three normative limitations of “fairness” as predictive parity Notions of fairness of this form have several key limitations: 1. They legitimize and perpetuate inequalities justified by “merit.” Where does inequality in M come from? 2. They are narrowly bracketed. Inequality in W in the algorithm, instead of some outcomes Y in a wider population. 3. Fairness-based perspectives focus on categories (protected groups) and ignore within-group inequality. 24 / 28
  • 33. Social welfare as an alternative framework • The framework of fairness / bias / discrimination contrasts with perspectives focused on consequences for social welfare. • Common presumption for most theories of justice: Normative statements about society are based on statements about individual welfare • Formally: • Individuals i = 1, . . . , n • Individual i’s welfare Yi • Social welfare as function of individuals’ welfare SWF = F(Y1, . . . , Yn). • Key points of contention: 1. Who is included among the individuals i? Who’s lives matter? 2. How to measure individual welfare Yi? 3. How to trade off welfare across individuals i? 25 / 28
  • 34. The impact on inequality or welfare as an alternative to fairness • Outcomes are determined by the potential outcome equation Y = W · Y1 + (1 − W) · Y0 . • The realized outcome distribution is given by pY,X(y, x) = h pY0|X(y, x) + w(x) · pY1|X(y, x) − pY0|X(y, x) i · pX(x). • What is the impact of w(·) on a statistic ν? ν = ν(pY,X). Examples: Variance, quantiles, between group inequality, social welfare. • Cf. Distributional decompositions in labor economics! 26 / 28
  • 35. When fairness and equality are in conflict • Fairness is about treating people of the same “merit” independently of their group membership. • Equality is about the (counterfactual / causal) consequences of an algorithm for the distribution of welfare of different people. Examples when they are in conflict: 1. Increased surveillance / better prediction algorithms: Lead to treatments more aligned with “merit” Good for fairness, bad for equality. 2. Affirmative action / compensatory interventions for pre-existing inequalities: Bad for fairness, good for equality. 27 / 28
  • 36. Conclusion • These two projects discuss settings showing the limitations of decision theory for understanding important current debates. • Other authors (in econometrics, economic theory, computer science, and elsewhere) are exploring related ideas! The road ahead: 1. Reconceptualize statistics as knowledge production and communication in a social context, involving diverse actors with divergent interests. Provide formal, prescriptive guidance for researchers based on this perspective. 2. Develop an understanding of algorithmic decision making / ML / AI based on the awareness that different people have different objectives, and control rights over data and algorithms make a difference. 28 / 28