SlideShare una empresa de Scribd logo
1 de 52
Descargar para leer sin conexión
Week 8:
Multinomial Logit Regression
Applied Statistical Analysis II
Jeffrey Ziegler, PhD
Assistant Professor in Political Science & Data Science
Trinity College Dublin
Spring 2023
Roadmap through Stats Land
Where we’ve been:
Over-arching goal: We’re learning how to make inferences
about a population from a sample
Before reading week: We learned how to assess the model fit
of logit regressions
Today we will learn how to:
Estimate and interpret linear models when our
outcome is categorical
1 51
Logit regression to multinomial regression
For a binary outcome we estimate a logit regression as...
logit[P(Yi = 1|Xi)] = ln
 P(Yi = 1|Xi)
1 − P(Yi = 1|Xi)

= β0 + β1Xi
Intercept (β0): When x = 0, the odds that Y=1 versus the
baseline Y = 0 are expected to be exp(β0)
Slope (β1): When x increases by one unit, the odds that Y=1
versus the baseline Y = 0 are expected to multiply by a
factor of exp(β1)
Review of logistic regression estimation
How do we alter this framework for an outcome with more than
two categories?
2 51
Multinomial response variable
Suppose we have a response variable that can take three
possible outcomes that are coded as 1, 2, 3
More generally, the outcome y is categorical and can take
values 1, 2, ..., k such that (k  2)
P(Y = 1) = p1, P(Y = 2) = p2, ..., P(Y = k) = pk
k
X
j=1
pj = 1
3 51
Multinomial regression: Estimation
So, let 1 be the baseline category
ln

pi2
pi1

= β02 + β12Xi
ln

pi3
pi1

= β03 + β13Xi
4 51
How to think about multiplying odds
Starting odds Ending odds Starting Pr[Y=1] Ending Pr[Y=1]
0.0001 0.0006 0.0001 0.0006
0.001 0.006 0.001 0.006
0.01 0.06 0.01 0.057
0.1 0.6 0.091 0.38
1. 6. 0.5 0.9
10. 60. 0.91 1.
100. 600. 0.99 1.
5 51
How to think about multiplying odds
As an additive change, not much of a difference between a
probability of 0.0001 and 0.0006 (it’s only 0.05%)
Also not much difference between 0.99 and 1 (only 1%)
Largest additive effect occurs when the odds equal 1
√
β
,
where the probability changes from 29% to 71% (↑ 42%)
6 51
Example: Expansion of religious groups
When and where do religious groups expand?
Religious marketplace: # of followers, competition from
other denominations
Political marketplace: Government policies around religion
Case study: Expansion of dioceses in Catholic Church (1900-2010)
Outcome: Diocese change in a given country (-, no ∆, +)
Predictors:
I % of population that is Catholic
I % of pop. that is Christian, but not Catholic (Protestant,
Evangelical, etc.)
I Government religiousity: Non-Christian, Christian,
Non-Catholic, Catholic
7 51
Variation in government religiousity
Venezuela
Spain
Portugal
Norway
Netherlands
Mexico
Italy
Ireland
Germany
France
Ecuador
Costa Rica
Colombia
Chile
Belgium
Austria
1
8
9
8
1
9
0
8
1
9
1
8
1
9
2
8
1
9
3
8
1
9
4
8
1
9
5
8
1
9
6
8
1
9
7
8
1
9
8
8
1
9
9
8
2
0
0
8
Year
Country
Non−Christian Christian, non−Catholic Catholic
8 51
Multinomial regression: Estimation in R
1 # import data
2 diocese_data − read . csv (  https : //raw . githubusercontent . com/ASDS−TCD/ S t a t s I I _Spring2023/
main/datasets/diocese_data . csv  , stringsAsFactors = F )
1 # run base multinomial l o g i t
2 multinom_model1 − multinom ( state3 ~ cathpct_ lag + hogCatholic_ lag +
3 propenpt_ lag + country ,
4 data = diocese_data )
Decrease Increase
% Catholic −0.069∗∗ −0.007
(0.025) (0.012)
% Protestant  Evangelical −0.035 0.033∗
(0.042) (0.015)
GovernmentCatholic 0.579 0.519∗
(0.504) (0.234)
GovernmentChristian,non−Catholic −148.089 −0.151
(0.646) (0.862)
Constant 2.128 −0.839
(2.226) (1.192)
Deviance 1896.567 1896.567
N 2808 2808
∗∗∗p  0.001, ∗∗p  0.01, ∗p  0.05
Now what?
9 51
Multinomial regression: Interpretation
For every one unit increase in X, the log-odds of Y = j vs.
Y = 1 increase by β1j
For every one unit increase in X, the odds of Y = j vs. Y = 1
multiply by a factor of exp(β1j)
ln
piIncrease
piNone
!
= −0.839 + 0.33X%Protestant + 0.519XGov0tCatholic − 0.151XGov0tChrisitan − 0.007X%Catholic
ln
piDecrease
piNone
!
= 2.128 − 0.35X%Protestant + 0.579XGov0tCatholic − 148.089XGov0tChrisitan − 0.069X%Catholic
10 51
Multinomial regression: Interpretation
1 # exponentiate coefficients
2 exp ( coef ( multinom_model1 ) [ , c ( 1 : 5 ) ] )
(Intercept) % Catholic Catholic Gov’t Christian, non-Catholic Gov’t % Protestant
Decrease 8.40 0.93 1.78 0.00 0.97
Increase 0.43 0.99 1.68 0.86 1.03
So, in a given country, there is an increase in the baseline
odds that the Church will expand by 1.68 times when a
government supports policies that the Church supports
11 51
Multinomial regression: Prediction
For j = 2, ..., k, we calculate the probability pij as
pij =
exp(β0j + β1jxi)
1 +
Pk
j=2 exp(β0j + β1jxi)
For baseline category (j = 1) we calculate probability (pi1) as
pi1 = 1 −
k
X
j=2
pij
We’ll use these probabilities to assign a category of the response
for each observation
12 51
Prediction in R: Creating data
1 # create hypothetical cases for predicted values
2 predict _data − data . frame ( hogCatholic_ lag = rep ( c ( Non− Christian  ,  Catholic  , 
Christian , non− Catholic  ) , each = 3) , cathpct_ lag = rep ( c (25 , 50 , 75) , 3) ,
3 propenpt_ lag = rep ( c (75 , 50 , 25) , 3) , country= Spain  )
hogCatholic_lag cathpct_lag propenpt_lag country
1 Non-Christian 25.00 75.00 Spain
2 Non-Christian 50.00 50.00 Spain
3 Non-Christian 75.00 25.00 Spain
4 Catholic 25.00 75.00 Spain
5 Catholic 50.00 50.00 Spain
6 Catholic 75.00 25.00 Spain
7 Christian, non-Catholic 25.00 75.00 Spain
8 Christian, non-Catholic 50.00 50.00 Spain
9 Christian, non-Catholic 75.00 25.00 Spain
13 51
Prediction in R: Extract predicted values
1 # store the predicted probabilities for each covariate combo
2 predicted_values − cbind ( predict _data , predict ( multinom_model1 , newdata = predict _data ,
type =  probs  , se = TRUE ) )
3
4 # calculate the mean probabilities within each level of government r e l i g i o u s i t y
5 by ( predicted_values [ , 5 : 7 ] , predicted_values$hogCatholic_lag , colMeans )
predicted_values$hogCatholic_lag: Catholic
None Decrease Increase
0.45543215 0.04051514 0.50405271
----------------------------------------------------
predicted_values$hogCatholic_lag: Christian, non-Catholic
None Decrease Increase
6.144487e-01 1.621855e-66 3.855513e-01
-----------------------------------------------------
predicted_values$hogCatholic_lag: Non-Christian
None Decrease Increase
0.56744365 0.03013509 0.40242126
14 51
Prediction in R: Plot
None
Decrease
Increase
3
0
4
0
5
0
6
0
7
0
0.0
0.2
0.4
0.6
0.8
0.0
0.2
0.4
0.6
0.8
0.0
0.2
0.4
0.6
0.8
% Catholic
Probability
Catholic Christian, non−Catholic Non−Christian
15 51
Accuracy of predictions
1
2 # see how well our predictions were :
3 addmargins ( table ( diocese_data$state3 , predict ( multinom_
model1 , type= class  ) ) )
None Decrease Increase Sum
None 2290.00 0.00 59.00 2349.00
Decrease 26.00 0.00 2.00 28.00
Increase 290.00 0.00 141.00 431.00
Sum 2606.00 0.00 202.00 2808.00
Thought for later: Not great prediction accuracy, maybe we want
a different model?
16 51
Segway: Model Fit  Assessment
For each category of the response, j:
Analyze a plot of the binned residuals vs. predicted
probabilities
Analyze a plot of the binned residuals vs. predictor
Look for any patterns in the residuals plots
For each categorical predictor variable, examine the average
residuals for each category of outcome
17 51
Calculating and plotting residuals
For response categories j and covariate patterns k, the Pearson residual is given by
rjk =
yjk − µ̂jk
q

var Yjk
=
yjk − µ̂jk
q
µ̂jk
where yjk is the observed count and µ̂jkthe expected count according to your fitted model
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
● ●
● ●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●●●●●●●●●●●●●●●●●
●
●
●
●
●
●●
●●●
●
●
●
●
●
●
●
●●●●
●
●
●
●●●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
0.00 0.05 0.10 0.15 0.20
−0.2
0.0
0.2
0.4
0.6
0.8
1.0
Pr(Y=Decrease)
Pearson
residuals
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
0.0 0.2 0.4 0.6 0.8
−0.5
0.0
0.5
1.0
Pr(Y=Increase)
Pearson
residuals
18 51
Potential Problems: Focusing on Coefficients
Can lead researchers to overlook significant effects that
affect their conclusions1
I Ex: Evaluate effect of information from events and messages
on individuals’ beliefs about war in Iraq (Gelpi 2017)2
Participants in experiment received different events and cues
Events treatment: subjects read a newspaper story with
either good news, bad news, or a control condition with a
news story that has no news events
Opinions treatment: newspaper story includes a confident
statement from President Bush, cautious statements, or none
Outcome: Opinion on a scale regarding whether (1) surge was a
success, (2) US would succeed in Iraq, and (3) should be a
timetable for withdrawal
1
Paolino, Philip. Predicted Probabilities and Inference with Multinomial Logit. Political Analysis 29.3 (2021):
416-421.
2
Gelpi, Christopher. The Surprising Robustness of Surprising Events: A Response to a Critique of “Performing on
Cue”. Journal of Conflict Resolution 61.8 (2017): 1816-1834.
19 51
Potential Problems: Focusing on Coefficients
Original findings: little evidence of any impact for elite
opinion at all in this experiment”
I Only coefficient for elite opinion that had a statistically
significant effect was among respondents who “strongly
approved” of Bush exposed to a cautious message from Bush
were more likely to “somewhat disapprove” of a timetable for
withdrawal
Overlooks a significant effect of message on attitudes of
respondents who “somewhat approved” of Bush
Predicted probs show: respondents had 0.177 (se=.073) lower
predicted probability of “strongly disapproving” and a .157
(se=.076) greater probability of “somewhat disapproving” of
a timetable than those not exposed to a cautious message
20 51
Potential Problems: Focusing on Coefficients
So, cautious message influenced respondents’ attitudes
about a timetable in both cases
I From this, we might instead conclude that elite opinion had
greater influence than events upon participants’ attitudes
toward withdrawal
21 51
Potential Problems: Interpretation Against Baseline
Sometimes you use specific baseline because RQ concerns
change in occurrence of one outcome rather than another
I If interested in relative odds, may seem on safer ground using
coefficients for inference
I But, should still calculate changes in predicted probabilities
of relevant outcomes to understand changes in underlying
probabilities
I A significant log odds ratio resulting from a large change in
probability of only one outcome may produce a substantively
different interpretation than one where both probabilities
change significantly
22 51
Potential Problems: Interpretation Against Baseline
Ex: Effect of trust upon acceptance of rumors in conflict zones
(Greenhill and Oppenheim 2017)3
Assess hypothesis that as distrust of implicated entity rises,
more likely it is that a rumor will be perceived as possibly or
definitely true” against a baseline of disbelief (663)
Given choice of baseline, effect of distrust or threat is
significant for 0 or only 1 response categories
When coefficients for distrust or threat are statistically
significant for both response categories, authors interpret
these instances as indicating higher odds of being in both
categories compared to baseline: “threat perception
increase[s] odds of being in both agnostic and receptive
categories” (668)
3
Greenhill, Kelly M., and Ben Oppenheim. Rumor has it: The adoption of unverified information in conflict zones.
International Studies Quarterly 61.3 (2017): 660-676.
23 51
Potential Problems: Interpretation Against Baseline
This interpretation of risk ratio, however, can be misleading
0
.2
.4
.6
.8
Predicted
Probability
Very Unlikely Very Likely
perceived likelihood of conflict incident in locality in next year
Deny Plausible Accept
Acceptance of Coup Rumors
24 51
Potential Problems: Interpretation Against Baseline
0
.2
.4
.6
.8
Predicted
Probability
Very Unlikely Very Likely
perceived likelihood of conflict incident in locality in next year
Deny Plausible Accept
Acceptance of Coup Rumors
Threat perceptions
increase both Pr(rumor
of coup) in Thailand as
being plausible and
being accepted
Although, little change
in Pr(accept rumor) at
levels of statistical
significance below
p  0.01
25 51
Potential Problems with MNL Coefficients
Interpretations of odds against a baseline often imply
significant coefficient = change in probabilities of both
alternatives
I But, change predicted probability of 1 alternative with a
significant coefficient may be no different than change in
predicted probability with a non significant coefficient
I Because we often rely upon significance levels for testing
hypotheses, demonstrates importance of examining changes
in predicted probabilities
Even when you’re interested only in effect of a covariate on
odds of one category against a baseline
26 51
Unordered to ordered multinomial regression
With ordinal data, let Yi = j if individual i chooses option j
j = 0, .., J, so we have J + 1 choices
Examples
I Political economy: Budgets (↓, no ∆, ↑)
I Medicine: Patient pain
I Biology: Animal extinction (Extinct, Endangered, Protected, No
regulation)
How should we model Pr(Yi = j) to get a model we can estimate?
Answer: Pr(Yi = j) will be areas under a probability density function (pdf)
I Normal pdf gives ordered probit, logistic gives ordered logit
27 51
Estimating Cutoffs: Intuition with logit
Binary dependent variable Y
= 0 or 1 (e.g., Male/Female)
Estimate
logit[P(Yi = 1|Xi)] =
ln

P(Yi=1|Xi)
1−P(Yi=1|Xi)

= β0 + β1Xi
Calculate
Xiβ = ŷ = β0 + β1Xi + ... for
each observation
For observation i, if ŷ  0.5
then predict Yi = 0; if ŷ  0
predict Yi = 1
28 51
Estimating Cutoffs for ordered logit
For instance, if b  Xiβ  a then predict Yi = µ6 − µ5
Interpretation of β1: ↑ x1 by one unit changes the probability of going to
next µj by β units
The impact of this on Pr(Y = 1) depends on your starting point
29 51
Example: Concern about COVID-19 in U.S.
How concerned are you that you or someone you know will be
infected with the coronavirus?
ABC News, web-based, nationally representative of U.S. (Beginning of
pandemic; April 15 - 16, 2020)
Political Hypothesis: Party ID predicts concern (Republicans less
concerned than Democrats)
Potential confounders:
Personal risk: Age
Personal hygiene: In the past week have you worn a face mask or face
covering when you’ve left your home, or not [Yes/No]
Level of potential local infection: Population density [Metro/rural]
Public health policies: Region [South, Northeast, Midwest, West]
30 51
COVID-19 concern over time, by party, region...
South MidWest NorthEast West
A
n
I
n
d
e
p
e
n
d
e
n
t
1
2
3
A
n
I
n
d
e
p
e
n
d
e
n
t
1
2
3
A
n
I
n
d
e
p
e
n
d
e
n
t
1
2
3
A
n
I
n
d
e
p
e
n
d
e
n
t
1
2
3
0
10
20
30
40
Party ID
Count
in
Region
Not concerned at all Not so concerned
Somewhat concerned Very concerned
31 51
Ordered Multinomial: Estimation in R
1 exp ( cbind (OR = coef ( ordered_ l o g i t ) , confint ( ordered_ l o g i t ) ) )
Did not leave home in the past week 1.052∗∗∗
(0.306)
Wear mask outside 1.271∗∗∗
(0.206)
Democrat 0.970∗∗∗
(0.208)
Republican −0.846∗∗∗
(0.215)
Metro area −0.274
(0.460)
MidWest −1.369∗
(0.571)
NorthEast 0.110
(0.712)
West −0.813
(0.758)
Metro area × MidWest 1.318∗
(0.627)
Metro area × NorthEast 0.077
(0.763)
Metro area × West 0.475
(0.793)
Deviance 1041.209
Num. obs. 514
∗∗∗p  0.001, ∗∗p  0.01, ∗p  0.05
What’s missing? Intercept and cutoffs! Don’t worry, we’ll come
back to this...
32 51
Ordered Multinomial: Interpretation
For now, how do we interpret the estimated coefficients? Similar
procedure to multinomial logit!
For every one unit increase in X, the log-odds of Y = j vs.
Y = j + 1 increase by β1
For every one unit increase in X, the odds of Y = j vs.
Y = j + 1 multiply by a factor of exp(β1)
ln
pij+1
pij
!
= 1.052 × No leave + 1.271 × Wear mask + 0.970 × Democrat + ...+
0.475 × Metro area × West
33 51
Ordered Multinomial: Interpretation
1 # get odds ratios and CIs
2 exp ( cbind (OR = coef ( ordered_ l o g i t ) , confint ( ordered_ l o g i t ) ) )
OR 2.5 % 97.5 %
Did not leave home in the past week 2.65 1.45 4.90
Wear mask outside 3.52 2.35 5.30
Democrat 2.65 1.76 4.02
Republican 0.43 0.28 0.66
Metro area 0.76 0.30 1.85
MidWest 0.25 0.08 0.77
NorthEast 1.12 0.28 4.59
West 0.44 0.10 1.99
Metro area×MidWest 3.74 1.10 12.98
Metro area×NorthEast 1.08 0.24 4.82
Metro area×West 1.61 0.33 7.67
Sanity X: For those who take protective measures (staying at home,
wearing a mask outside), odds of being more concerned about COVID are
2.6 to 3.5× higher compared to those who do not wear masks outside,
holding constant all other variables
34 51
Ordered Multinomial: Further Interpretation
OR 2.5 % 97.5 %
Did not leave home in the past week 2.65 1.45 4.90
Wear mask outside 3.52 2.35 5.30
Democrat 2.65 1.76 4.02
Republican 0.43 0.28 0.66
Metro area 0.76 0.30 1.85
MidWest 0.25 0.08 0.77
NorthEast 1.12 0.28 4.59
West 0.44 0.10 1.99
Metro area×MidWest 3.74 1.10 12.98
Metro area×NorthEast 1.08 0.24 4.82
Metro area×West 1.61 0.33 7.67
So, for a Democrat respondent in a given region, the odds of being more
concerned about COVID are 2.65× higher compared to an Independent
respondent, holding constant all other variables
Geography: Population density and region
35 51
Model assumptions: Proportional odds
Major assumption underlying ordinal logit ( ordinal probit)
regression is that relationship between each pair of outcome
groups is equal
In other words, ordinal logistic regression assumes
coefficients that describe relationship between lowest
versus all higher categories of response variable are same as
those that describe relationship between next lowest
category and all higher categories, etc.
Called proportional odds assumption or parallel regression
assumption
Because relationship between all pairs of groups is same,
there is only one set of coefficients
36 51
Checking the parallel lines assumption
Estimate and compare simplified model:
1 # run l o g i t regressions for a l l outcome categories
2 for ( i in 1 : length ( unique ( covid_data$covid_concern ) ) ) {
3 assign ( paste (  l o g i t _model , i , sep=  ) , glm ( i f e l s e ( covid_concern==unique ( covid_data$
covid_concern ) [ i ] , 1 , 0) ~ mask + partyID , data = covid_data ) , envir =globalenv ( ) )
4 }
Not concerned at all Not so concerned Somewhat concerned Very concerned
Did not leave home in the past week −0.014 −0.169∗∗ −0.007 0.190∗∗
(0.029) (0.053) (0.076) (0.072)
Wear mask outside −0.073∗∗∗ −0.106∗∗ −0.085 0.264∗∗∗
(0.019) (0.036) (0.051) (0.048)
Democrat −0.000 −0.121∗∗∗ −0.101∗ 0.222∗∗∗
(0.019) (0.035) (0.050) (0.047)
Republican 0.022 0.139∗∗∗ −0.026 −0.134∗∗
(0.020) (0.038) (0.054) (0.051)
(Intercept) 0.078∗∗∗ 0.242∗∗∗ 0.483∗∗∗ 0.197∗∗∗
(0.018) (0.034) (0.049) (0.046)
Deviance 16.729 58.186 120.155 106.944
Num. obs. 514 514 514 514
∗∗∗p  0.001, ∗∗p  0.01, ∗p  0.05
Doesn’t look like parallel odds holds, maybe not ordered
categories (or wrong categories)?
37 51
Ordered Multinomial: Prediction
Remember those cutoffs?! This is where they’re useful
Assume Y has more than two ordered categories (for instance, low,
medium, high)
We now need two cut-points to divide the curve into three sections
We’ll estimate these as µ1 and µ2 using MLE
38 51
Ordered Multinomial: Prediction
If Xiβ  µ1 then
predict Yi = low
If µ1  Xiβ  µ2 then
predict Yi = medium
If Xiβ  µ2 then
predict Yi = high
β ∆ in x may change the prediction on Y, or it may not!
39 51
Prediction in R: Creating data
1 # create fake data to use for prediction
2 predict _data − data . frame (mask = rep ( c ( No ,  Yes  ) , 3) ,
3 partyID = rep ( c ( An Independent  , A Democrat , A Republican  ) , each = 2) )
Wear mask party ID
1 No An Independent
2 Yes An Independent
3 No A Democrat
4 Yes A Democrat
5 No A Republican
6 Yes A Republican
40 51
Prediction in R: Extract predicted values
1 # add in predictions for each observation in fake data
2 plot _data − melt ( cbind ( predict _data , predict ( ordered_ l o g i t _slim , predict _data , type = 
probs  ) ) , id . vars = c ( mask ,  partyID  ) ,
3 variable . name =  Level  , value . name= Probability  )
Wear mask party ID Level of concern Pr(Y = j)
1 No An Independent Not concerned at all 0.06
2 Yes An Independent Not concerned at all 0.02
3 No A Democrat Not concerned at all 0.03
4 Yes A Democrat Not concerned at all 0.01
.
.
.
.
.
.
.
.
.
.
.
.
24 Yes A Republican Very concerned 0.27
41 51
Prediction in R: Plot
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
Not concerned at all Not so concerned Somewhat concerned Very concerned
A
D
e
m
o
c
r
a
t
A
R
e
p
u
b
l
i
c
a
n
A
n
I
n
d
e
p
e
n
d
e
n
t
A
D
e
m
o
c
r
a
t
A
R
e
p
u
b
l
i
c
a
n
A
n
I
n
d
e
p
e
n
d
e
n
t
A
D
e
m
o
c
r
a
t
A
R
e
p
u
b
l
i
c
a
n
A
n
I
n
d
e
p
e
n
d
e
n
t
A
D
e
m
o
c
r
a
t
A
R
e
p
u
b
l
i
c
a
n
A
n
I
n
d
e
p
e
n
d
e
n
t
0.0
0.2
0.4
0.6
Party ID
Pr(Y=j)
Wear Mask ● ●
No Yes
42 51
Accuracy of predictions
1 # see how well our predictions were :
2 addmargins ( table ( covid_data$covid_concern , predict ( ordered_
l o g i t _slim , type=  class  ) ) )
Not concerned at all Not so concerned Somewhat concerned Very concerned Sum
Not concerned at all 0 6 7 5 18
Not so concerned 0 18 38 21 77
Somewhat concerned 0 11 102 87 200
Very concerned 0 7 54 158 219
Sum 0 42 201 271 514
43 51
Another Ex: Social Trust and Political Interest
Is paying attention to politics associated with less trust in
others?
H1: The more people pay attention to negative political
messages the less they trust others
1 # load in example 2 data :
2 # ESS p o l i t i c a l and social trust
3 # clean up data
4 # need to have registered email
5 set_email ( me@tcd. ie  )
6 # import a l l rounds from IE
7 ESS_data − import_ a l l _cntrounds ( country =  Ireland  )
44 51
Social Trust and Political Interest: Our Model
Outcome:
ppltrst: Most people can be trusted or you can’t be too
careful
I Values: Ordinal 0-11
Predictors:
polintr: How interested in politics
I Values: 1-4 (not at all, not really, somewhat, quite)
essround: ESS round (time)
45 51
Estimate model of trust and political interest
1 # run ordered l o g i t
2 ordered_ trust − polr ( ppltrst ~ polintr + essround , data =
merged_ESS , Hess=T )
3
4 # get odds ratios and CIs
5 exp ( cbind (OR = coef ( ordered_ trust ) , confint ( ordered_ trust ) )
)
OR 2.5 % 97.5 %
polintr2 0.95 0.88 1.04
polintr3 0.76 0.69 0.83
polintr4 0.58 0.53 0.64
essround2 1.35 1.21 1.50
essround3 0.89 0.80 0.99
essround4 0.91 0.81 1.014
essround5 0.78 0.70 0.86
essround6 0.75 0.68 0.83
essround7 0.80 0.72 0.89
essround8 2.29 2.07 2.54
essround9 2.05 1.84 2.28
46 51
Predicted probabilities with CIs
1 # get predicted probs for each category
2 # for person with high p o l i t i c a l interest ( 4 )
3 # from last wave
4 # use l i b r a r y ( glm . predict to get CIs )
5 basepredict . polr ( ordered_ trust , values = c ( rep (0 ,2) , 1 , rep
( 0 , 7 ) , 1 ) )
¯
ŷ 2.5% 97.5%
0 0.021 0.018 0.023
1 0.027 0.025 0.030
2 0.040 0.037 0.044
3 0.064 0.059 0.069
4 0.077 0.071 0.082
5 0.151 0.144 0.158
6 0.144 0.139 0.149
7 0.176 0.169 0.182
8 0.180 0.169 0.191
9 0.076 0.070 0.083
10 0.035 0.032 0.039
11 0.009 0.007 0.011
47 51
Change in Predicted probabilities with CIs
1 # see how predicted probs change between two cases
2 # low interest ( 1 ) to high ( 4 ) for wave 1
3 dc ( ordered_ trust , values1 = c ( rep (0 ,3) , 1 , rep ( 0 , 7 ) ) ,
4 values2 = c ( rep (0 ,2) , 1 , 1 , rep ( 0 , 7 ) ) )
¯
∆ 2.5% 97.5%
0 -0.013 -0.015 -0.010
1 -0.016 -0.018 -0.013
2 -0.021 -0.024 -0.017
3 -0.028 -0.032 -0.023
4 -0.025 -0.029 -0.021
5 -0.028 -0.033 -0.022
6 -0.002 -0.005 0.001
7 0.026 0.021 0.030
8 0.055 0.046 0.064
9 0.031 0.025 0.037
10 0.016 0.013 0.019
11 0.004 0.003 0.005
48 51
Wrap Up
In this lesson, we went over how to...
Extended logit regression for outcomes that have more than
one category
Extended multinomial logit regression for outcomes are
ordered categories
After reading week, we’ll talk about event  count outcome
variables
49 51
Review: Logit regression estimation
We want to think about how to take our set up from a binary logit
and apply it to multiple outcome categories
Without any loss of generality, index the Xi so that Xk is the
variable and we’re still interested in taking exp of βk
Fixing the values of each covariate and varying Xk by a small
amount yields some δ
log(p(..., Xk + δ)) − log(p(..., Xk)) = βkδ
Thus, βk is marginal change in log odds with respect to Xk
Back to slides
50 51
Review: Logit regression estimation
To recover exp(βk), we must set δ=1 and exponentiate left hand
side:
exp(βk) = exp(βk) × δ (1)
= exp(βk) (2)
= exp(log(p(..., Xk + δ)) − log(p(..., Xk))) (3)
=
p(..., Xk + 1)
p(..., Xk)
(4)
(5)
Back to slides
51 / 51

Más contenido relacionado

Último

Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfSherif Taha
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docxPoojaSen20
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingTechSoup
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.pptRamjanShidvankar
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.christianmathematics
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17Celine George
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and ModificationsMJDuyan
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
Magic bus Group work1and 2 (Team 3).pptx
Magic bus Group work1and 2 (Team 3).pptxMagic bus Group work1and 2 (Team 3).pptx
Magic bus Group work1and 2 (Team 3).pptxdhanalakshmis0310
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxVishalSingh1417
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhikauryashika82
 
Third Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxThird Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxAmita Gupta
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 

Último (20)

Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdf
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and Modifications
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
Magic bus Group work1and 2 (Team 3).pptx
Magic bus Group work1and 2 (Team 3).pptxMagic bus Group work1and 2 (Team 3).pptx
Magic bus Group work1and 2 (Team 3).pptx
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Third Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxThird Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptx
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 

Destacado

2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by HubspotMarius Sescu
 
Everything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTEverything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTExpeed Software
 
Product Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsProduct Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsPixeldarts
 
How Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthHow Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthThinkNow
 
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfAI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfmarketingartwork
 
PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024Neil Kimberley
 
Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)contently
 
How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024Albert Qian
 
Social Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsSocial Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsKurio // The Social Media Age(ncy)
 
Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Search Engine Journal
 
5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summarySpeakerHub
 
ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd Clark Boyd
 
Getting into the tech field. what next
Getting into the tech field. what next Getting into the tech field. what next
Getting into the tech field. what next Tessa Mero
 
Google's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentGoogle's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentLily Ray
 
Time Management & Productivity - Best Practices
Time Management & Productivity -  Best PracticesTime Management & Productivity -  Best Practices
Time Management & Productivity - Best PracticesVit Horky
 
The six step guide to practical project management
The six step guide to practical project managementThe six step guide to practical project management
The six step guide to practical project managementMindGenius
 
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...RachelPearson36
 

Destacado (20)

2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot2024 State of Marketing Report – by Hubspot
2024 State of Marketing Report – by Hubspot
 
Everything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPTEverything You Need To Know About ChatGPT
Everything You Need To Know About ChatGPT
 
Product Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage EngineeringsProduct Design Trends in 2024 | Teenage Engineerings
Product Design Trends in 2024 | Teenage Engineerings
 
How Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental HealthHow Race, Age and Gender Shape Attitudes Towards Mental Health
How Race, Age and Gender Shape Attitudes Towards Mental Health
 
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdfAI Trends in Creative Operations 2024 by Artwork Flow.pdf
AI Trends in Creative Operations 2024 by Artwork Flow.pdf
 
Skeleton Culture Code
Skeleton Culture CodeSkeleton Culture Code
Skeleton Culture Code
 
PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024PEPSICO Presentation to CAGNY Conference Feb 2024
PEPSICO Presentation to CAGNY Conference Feb 2024
 
Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)Content Methodology: A Best Practices Report (Webinar)
Content Methodology: A Best Practices Report (Webinar)
 
How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024How to Prepare For a Successful Job Search for 2024
How to Prepare For a Successful Job Search for 2024
 
Social Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie InsightsSocial Media Marketing Trends 2024 // The Global Indie Insights
Social Media Marketing Trends 2024 // The Global Indie Insights
 
Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024Trends In Paid Search: Navigating The Digital Landscape In 2024
Trends In Paid Search: Navigating The Digital Landscape In 2024
 
5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary5 Public speaking tips from TED - Visualized summary
5 Public speaking tips from TED - Visualized summary
 
ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd ChatGPT and the Future of Work - Clark Boyd
ChatGPT and the Future of Work - Clark Boyd
 
Getting into the tech field. what next
Getting into the tech field. what next Getting into the tech field. what next
Getting into the tech field. what next
 
Google's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search IntentGoogle's Just Not That Into You: Understanding Core Updates & Search Intent
Google's Just Not That Into You: Understanding Core Updates & Search Intent
 
How to have difficult conversations
How to have difficult conversations How to have difficult conversations
How to have difficult conversations
 
Introduction to Data Science
Introduction to Data ScienceIntroduction to Data Science
Introduction to Data Science
 
Time Management & Productivity - Best Practices
Time Management & Productivity -  Best PracticesTime Management & Productivity -  Best Practices
Time Management & Productivity - Best Practices
 
The six step guide to practical project management
The six step guide to practical project managementThe six step guide to practical project management
The six step guide to practical project management
 
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
Beginners Guide to TikTok for Search - Rachel Pearson - We are Tilt __ Bright...
 

8_multinomial_printable.pdf

  • 1. Week 8: Multinomial Logit Regression Applied Statistical Analysis II Jeffrey Ziegler, PhD Assistant Professor in Political Science & Data Science Trinity College Dublin Spring 2023
  • 2. Roadmap through Stats Land Where we’ve been: Over-arching goal: We’re learning how to make inferences about a population from a sample Before reading week: We learned how to assess the model fit of logit regressions Today we will learn how to: Estimate and interpret linear models when our outcome is categorical 1 51
  • 3. Logit regression to multinomial regression For a binary outcome we estimate a logit regression as... logit[P(Yi = 1|Xi)] = ln P(Yi = 1|Xi) 1 − P(Yi = 1|Xi) = β0 + β1Xi Intercept (β0): When x = 0, the odds that Y=1 versus the baseline Y = 0 are expected to be exp(β0) Slope (β1): When x increases by one unit, the odds that Y=1 versus the baseline Y = 0 are expected to multiply by a factor of exp(β1) Review of logistic regression estimation How do we alter this framework for an outcome with more than two categories? 2 51
  • 4. Multinomial response variable Suppose we have a response variable that can take three possible outcomes that are coded as 1, 2, 3 More generally, the outcome y is categorical and can take values 1, 2, ..., k such that (k 2) P(Y = 1) = p1, P(Y = 2) = p2, ..., P(Y = k) = pk k X j=1 pj = 1 3 51
  • 5. Multinomial regression: Estimation So, let 1 be the baseline category ln pi2 pi1 = β02 + β12Xi ln pi3 pi1 = β03 + β13Xi 4 51
  • 6. How to think about multiplying odds Starting odds Ending odds Starting Pr[Y=1] Ending Pr[Y=1] 0.0001 0.0006 0.0001 0.0006 0.001 0.006 0.001 0.006 0.01 0.06 0.01 0.057 0.1 0.6 0.091 0.38 1. 6. 0.5 0.9 10. 60. 0.91 1. 100. 600. 0.99 1. 5 51
  • 7. How to think about multiplying odds As an additive change, not much of a difference between a probability of 0.0001 and 0.0006 (it’s only 0.05%) Also not much difference between 0.99 and 1 (only 1%) Largest additive effect occurs when the odds equal 1 √ β , where the probability changes from 29% to 71% (↑ 42%) 6 51
  • 8. Example: Expansion of religious groups When and where do religious groups expand? Religious marketplace: # of followers, competition from other denominations Political marketplace: Government policies around religion Case study: Expansion of dioceses in Catholic Church (1900-2010) Outcome: Diocese change in a given country (-, no ∆, +) Predictors: I % of population that is Catholic I % of pop. that is Christian, but not Catholic (Protestant, Evangelical, etc.) I Government religiousity: Non-Christian, Christian, Non-Catholic, Catholic 7 51
  • 9. Variation in government religiousity Venezuela Spain Portugal Norway Netherlands Mexico Italy Ireland Germany France Ecuador Costa Rica Colombia Chile Belgium Austria 1 8 9 8 1 9 0 8 1 9 1 8 1 9 2 8 1 9 3 8 1 9 4 8 1 9 5 8 1 9 6 8 1 9 7 8 1 9 8 8 1 9 9 8 2 0 0 8 Year Country Non−Christian Christian, non−Catholic Catholic 8 51
  • 10. Multinomial regression: Estimation in R 1 # import data 2 diocese_data − read . csv ( https : //raw . githubusercontent . com/ASDS−TCD/ S t a t s I I _Spring2023/ main/datasets/diocese_data . csv , stringsAsFactors = F ) 1 # run base multinomial l o g i t 2 multinom_model1 − multinom ( state3 ~ cathpct_ lag + hogCatholic_ lag + 3 propenpt_ lag + country , 4 data = diocese_data ) Decrease Increase % Catholic −0.069∗∗ −0.007 (0.025) (0.012) % Protestant Evangelical −0.035 0.033∗ (0.042) (0.015) GovernmentCatholic 0.579 0.519∗ (0.504) (0.234) GovernmentChristian,non−Catholic −148.089 −0.151 (0.646) (0.862) Constant 2.128 −0.839 (2.226) (1.192) Deviance 1896.567 1896.567 N 2808 2808 ∗∗∗p 0.001, ∗∗p 0.01, ∗p 0.05 Now what? 9 51
  • 11. Multinomial regression: Interpretation For every one unit increase in X, the log-odds of Y = j vs. Y = 1 increase by β1j For every one unit increase in X, the odds of Y = j vs. Y = 1 multiply by a factor of exp(β1j) ln piIncrease piNone ! = −0.839 + 0.33X%Protestant + 0.519XGov0tCatholic − 0.151XGov0tChrisitan − 0.007X%Catholic ln piDecrease piNone ! = 2.128 − 0.35X%Protestant + 0.579XGov0tCatholic − 148.089XGov0tChrisitan − 0.069X%Catholic 10 51
  • 12. Multinomial regression: Interpretation 1 # exponentiate coefficients 2 exp ( coef ( multinom_model1 ) [ , c ( 1 : 5 ) ] ) (Intercept) % Catholic Catholic Gov’t Christian, non-Catholic Gov’t % Protestant Decrease 8.40 0.93 1.78 0.00 0.97 Increase 0.43 0.99 1.68 0.86 1.03 So, in a given country, there is an increase in the baseline odds that the Church will expand by 1.68 times when a government supports policies that the Church supports 11 51
  • 13. Multinomial regression: Prediction For j = 2, ..., k, we calculate the probability pij as pij = exp(β0j + β1jxi) 1 + Pk j=2 exp(β0j + β1jxi) For baseline category (j = 1) we calculate probability (pi1) as pi1 = 1 − k X j=2 pij We’ll use these probabilities to assign a category of the response for each observation 12 51
  • 14. Prediction in R: Creating data 1 # create hypothetical cases for predicted values 2 predict _data − data . frame ( hogCatholic_ lag = rep ( c ( Non− Christian , Catholic , Christian , non− Catholic ) , each = 3) , cathpct_ lag = rep ( c (25 , 50 , 75) , 3) , 3 propenpt_ lag = rep ( c (75 , 50 , 25) , 3) , country= Spain ) hogCatholic_lag cathpct_lag propenpt_lag country 1 Non-Christian 25.00 75.00 Spain 2 Non-Christian 50.00 50.00 Spain 3 Non-Christian 75.00 25.00 Spain 4 Catholic 25.00 75.00 Spain 5 Catholic 50.00 50.00 Spain 6 Catholic 75.00 25.00 Spain 7 Christian, non-Catholic 25.00 75.00 Spain 8 Christian, non-Catholic 50.00 50.00 Spain 9 Christian, non-Catholic 75.00 25.00 Spain 13 51
  • 15. Prediction in R: Extract predicted values 1 # store the predicted probabilities for each covariate combo 2 predicted_values − cbind ( predict _data , predict ( multinom_model1 , newdata = predict _data , type = probs , se = TRUE ) ) 3 4 # calculate the mean probabilities within each level of government r e l i g i o u s i t y 5 by ( predicted_values [ , 5 : 7 ] , predicted_values$hogCatholic_lag , colMeans ) predicted_values$hogCatholic_lag: Catholic None Decrease Increase 0.45543215 0.04051514 0.50405271 ---------------------------------------------------- predicted_values$hogCatholic_lag: Christian, non-Catholic None Decrease Increase 6.144487e-01 1.621855e-66 3.855513e-01 ----------------------------------------------------- predicted_values$hogCatholic_lag: Non-Christian None Decrease Increase 0.56744365 0.03013509 0.40242126 14 51
  • 16. Prediction in R: Plot None Decrease Increase 3 0 4 0 5 0 6 0 7 0 0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8 % Catholic Probability Catholic Christian, non−Catholic Non−Christian 15 51
  • 17. Accuracy of predictions 1 2 # see how well our predictions were : 3 addmargins ( table ( diocese_data$state3 , predict ( multinom_ model1 , type= class ) ) ) None Decrease Increase Sum None 2290.00 0.00 59.00 2349.00 Decrease 26.00 0.00 2.00 28.00 Increase 290.00 0.00 141.00 431.00 Sum 2606.00 0.00 202.00 2808.00 Thought for later: Not great prediction accuracy, maybe we want a different model? 16 51
  • 18. Segway: Model Fit Assessment For each category of the response, j: Analyze a plot of the binned residuals vs. predicted probabilities Analyze a plot of the binned residuals vs. predictor Look for any patterns in the residuals plots For each categorical predictor variable, examine the average residuals for each category of outcome 17 51
  • 19. Calculating and plotting residuals For response categories j and covariate patterns k, the Pearson residual is given by rjk = yjk − µ̂jk q var Yjk = yjk − µ̂jk q µ̂jk where yjk is the observed count and µ̂jkthe expected count according to your fitted model ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●●●●●●●●●●●●●●●●● ● ● ● ● ● ●● ●●● ● ● ● ● ● ● ● ●●●● ● ● ● ●●●●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0.00 0.05 0.10 0.15 0.20 −0.2 0.0 0.2 0.4 0.6 0.8 1.0 Pr(Y=Decrease) Pearson residuals ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0.0 0.2 0.4 0.6 0.8 −0.5 0.0 0.5 1.0 Pr(Y=Increase) Pearson residuals 18 51
  • 20. Potential Problems: Focusing on Coefficients Can lead researchers to overlook significant effects that affect their conclusions1 I Ex: Evaluate effect of information from events and messages on individuals’ beliefs about war in Iraq (Gelpi 2017)2 Participants in experiment received different events and cues Events treatment: subjects read a newspaper story with either good news, bad news, or a control condition with a news story that has no news events Opinions treatment: newspaper story includes a confident statement from President Bush, cautious statements, or none Outcome: Opinion on a scale regarding whether (1) surge was a success, (2) US would succeed in Iraq, and (3) should be a timetable for withdrawal 1 Paolino, Philip. Predicted Probabilities and Inference with Multinomial Logit. Political Analysis 29.3 (2021): 416-421. 2 Gelpi, Christopher. The Surprising Robustness of Surprising Events: A Response to a Critique of “Performing on Cue”. Journal of Conflict Resolution 61.8 (2017): 1816-1834. 19 51
  • 21. Potential Problems: Focusing on Coefficients Original findings: little evidence of any impact for elite opinion at all in this experiment” I Only coefficient for elite opinion that had a statistically significant effect was among respondents who “strongly approved” of Bush exposed to a cautious message from Bush were more likely to “somewhat disapprove” of a timetable for withdrawal Overlooks a significant effect of message on attitudes of respondents who “somewhat approved” of Bush Predicted probs show: respondents had 0.177 (se=.073) lower predicted probability of “strongly disapproving” and a .157 (se=.076) greater probability of “somewhat disapproving” of a timetable than those not exposed to a cautious message 20 51
  • 22. Potential Problems: Focusing on Coefficients So, cautious message influenced respondents’ attitudes about a timetable in both cases I From this, we might instead conclude that elite opinion had greater influence than events upon participants’ attitudes toward withdrawal 21 51
  • 23. Potential Problems: Interpretation Against Baseline Sometimes you use specific baseline because RQ concerns change in occurrence of one outcome rather than another I If interested in relative odds, may seem on safer ground using coefficients for inference I But, should still calculate changes in predicted probabilities of relevant outcomes to understand changes in underlying probabilities I A significant log odds ratio resulting from a large change in probability of only one outcome may produce a substantively different interpretation than one where both probabilities change significantly 22 51
  • 24. Potential Problems: Interpretation Against Baseline Ex: Effect of trust upon acceptance of rumors in conflict zones (Greenhill and Oppenheim 2017)3 Assess hypothesis that as distrust of implicated entity rises, more likely it is that a rumor will be perceived as possibly or definitely true” against a baseline of disbelief (663) Given choice of baseline, effect of distrust or threat is significant for 0 or only 1 response categories When coefficients for distrust or threat are statistically significant for both response categories, authors interpret these instances as indicating higher odds of being in both categories compared to baseline: “threat perception increase[s] odds of being in both agnostic and receptive categories” (668) 3 Greenhill, Kelly M., and Ben Oppenheim. Rumor has it: The adoption of unverified information in conflict zones. International Studies Quarterly 61.3 (2017): 660-676. 23 51
  • 25. Potential Problems: Interpretation Against Baseline This interpretation of risk ratio, however, can be misleading 0 .2 .4 .6 .8 Predicted Probability Very Unlikely Very Likely perceived likelihood of conflict incident in locality in next year Deny Plausible Accept Acceptance of Coup Rumors 24 51
  • 26. Potential Problems: Interpretation Against Baseline 0 .2 .4 .6 .8 Predicted Probability Very Unlikely Very Likely perceived likelihood of conflict incident in locality in next year Deny Plausible Accept Acceptance of Coup Rumors Threat perceptions increase both Pr(rumor of coup) in Thailand as being plausible and being accepted Although, little change in Pr(accept rumor) at levels of statistical significance below p 0.01 25 51
  • 27. Potential Problems with MNL Coefficients Interpretations of odds against a baseline often imply significant coefficient = change in probabilities of both alternatives I But, change predicted probability of 1 alternative with a significant coefficient may be no different than change in predicted probability with a non significant coefficient I Because we often rely upon significance levels for testing hypotheses, demonstrates importance of examining changes in predicted probabilities Even when you’re interested only in effect of a covariate on odds of one category against a baseline 26 51
  • 28. Unordered to ordered multinomial regression With ordinal data, let Yi = j if individual i chooses option j j = 0, .., J, so we have J + 1 choices Examples I Political economy: Budgets (↓, no ∆, ↑) I Medicine: Patient pain I Biology: Animal extinction (Extinct, Endangered, Protected, No regulation) How should we model Pr(Yi = j) to get a model we can estimate? Answer: Pr(Yi = j) will be areas under a probability density function (pdf) I Normal pdf gives ordered probit, logistic gives ordered logit 27 51
  • 29. Estimating Cutoffs: Intuition with logit Binary dependent variable Y = 0 or 1 (e.g., Male/Female) Estimate logit[P(Yi = 1|Xi)] = ln P(Yi=1|Xi) 1−P(Yi=1|Xi) = β0 + β1Xi Calculate Xiβ = ŷ = β0 + β1Xi + ... for each observation For observation i, if ŷ 0.5 then predict Yi = 0; if ŷ 0 predict Yi = 1 28 51
  • 30. Estimating Cutoffs for ordered logit For instance, if b Xiβ a then predict Yi = µ6 − µ5 Interpretation of β1: ↑ x1 by one unit changes the probability of going to next µj by β units The impact of this on Pr(Y = 1) depends on your starting point 29 51
  • 31. Example: Concern about COVID-19 in U.S. How concerned are you that you or someone you know will be infected with the coronavirus? ABC News, web-based, nationally representative of U.S. (Beginning of pandemic; April 15 - 16, 2020) Political Hypothesis: Party ID predicts concern (Republicans less concerned than Democrats) Potential confounders: Personal risk: Age Personal hygiene: In the past week have you worn a face mask or face covering when you’ve left your home, or not [Yes/No] Level of potential local infection: Population density [Metro/rural] Public health policies: Region [South, Northeast, Midwest, West] 30 51
  • 32. COVID-19 concern over time, by party, region... South MidWest NorthEast West A n I n d e p e n d e n t 1 2 3 A n I n d e p e n d e n t 1 2 3 A n I n d e p e n d e n t 1 2 3 A n I n d e p e n d e n t 1 2 3 0 10 20 30 40 Party ID Count in Region Not concerned at all Not so concerned Somewhat concerned Very concerned 31 51
  • 33. Ordered Multinomial: Estimation in R 1 exp ( cbind (OR = coef ( ordered_ l o g i t ) , confint ( ordered_ l o g i t ) ) ) Did not leave home in the past week 1.052∗∗∗ (0.306) Wear mask outside 1.271∗∗∗ (0.206) Democrat 0.970∗∗∗ (0.208) Republican −0.846∗∗∗ (0.215) Metro area −0.274 (0.460) MidWest −1.369∗ (0.571) NorthEast 0.110 (0.712) West −0.813 (0.758) Metro area × MidWest 1.318∗ (0.627) Metro area × NorthEast 0.077 (0.763) Metro area × West 0.475 (0.793) Deviance 1041.209 Num. obs. 514 ∗∗∗p 0.001, ∗∗p 0.01, ∗p 0.05 What’s missing? Intercept and cutoffs! Don’t worry, we’ll come back to this... 32 51
  • 34. Ordered Multinomial: Interpretation For now, how do we interpret the estimated coefficients? Similar procedure to multinomial logit! For every one unit increase in X, the log-odds of Y = j vs. Y = j + 1 increase by β1 For every one unit increase in X, the odds of Y = j vs. Y = j + 1 multiply by a factor of exp(β1) ln pij+1 pij ! = 1.052 × No leave + 1.271 × Wear mask + 0.970 × Democrat + ...+ 0.475 × Metro area × West 33 51
  • 35. Ordered Multinomial: Interpretation 1 # get odds ratios and CIs 2 exp ( cbind (OR = coef ( ordered_ l o g i t ) , confint ( ordered_ l o g i t ) ) ) OR 2.5 % 97.5 % Did not leave home in the past week 2.65 1.45 4.90 Wear mask outside 3.52 2.35 5.30 Democrat 2.65 1.76 4.02 Republican 0.43 0.28 0.66 Metro area 0.76 0.30 1.85 MidWest 0.25 0.08 0.77 NorthEast 1.12 0.28 4.59 West 0.44 0.10 1.99 Metro area×MidWest 3.74 1.10 12.98 Metro area×NorthEast 1.08 0.24 4.82 Metro area×West 1.61 0.33 7.67 Sanity X: For those who take protective measures (staying at home, wearing a mask outside), odds of being more concerned about COVID are 2.6 to 3.5× higher compared to those who do not wear masks outside, holding constant all other variables 34 51
  • 36. Ordered Multinomial: Further Interpretation OR 2.5 % 97.5 % Did not leave home in the past week 2.65 1.45 4.90 Wear mask outside 3.52 2.35 5.30 Democrat 2.65 1.76 4.02 Republican 0.43 0.28 0.66 Metro area 0.76 0.30 1.85 MidWest 0.25 0.08 0.77 NorthEast 1.12 0.28 4.59 West 0.44 0.10 1.99 Metro area×MidWest 3.74 1.10 12.98 Metro area×NorthEast 1.08 0.24 4.82 Metro area×West 1.61 0.33 7.67 So, for a Democrat respondent in a given region, the odds of being more concerned about COVID are 2.65× higher compared to an Independent respondent, holding constant all other variables Geography: Population density and region 35 51
  • 37. Model assumptions: Proportional odds Major assumption underlying ordinal logit ( ordinal probit) regression is that relationship between each pair of outcome groups is equal In other words, ordinal logistic regression assumes coefficients that describe relationship between lowest versus all higher categories of response variable are same as those that describe relationship between next lowest category and all higher categories, etc. Called proportional odds assumption or parallel regression assumption Because relationship between all pairs of groups is same, there is only one set of coefficients 36 51
  • 38. Checking the parallel lines assumption Estimate and compare simplified model: 1 # run l o g i t regressions for a l l outcome categories 2 for ( i in 1 : length ( unique ( covid_data$covid_concern ) ) ) { 3 assign ( paste ( l o g i t _model , i , sep= ) , glm ( i f e l s e ( covid_concern==unique ( covid_data$ covid_concern ) [ i ] , 1 , 0) ~ mask + partyID , data = covid_data ) , envir =globalenv ( ) ) 4 } Not concerned at all Not so concerned Somewhat concerned Very concerned Did not leave home in the past week −0.014 −0.169∗∗ −0.007 0.190∗∗ (0.029) (0.053) (0.076) (0.072) Wear mask outside −0.073∗∗∗ −0.106∗∗ −0.085 0.264∗∗∗ (0.019) (0.036) (0.051) (0.048) Democrat −0.000 −0.121∗∗∗ −0.101∗ 0.222∗∗∗ (0.019) (0.035) (0.050) (0.047) Republican 0.022 0.139∗∗∗ −0.026 −0.134∗∗ (0.020) (0.038) (0.054) (0.051) (Intercept) 0.078∗∗∗ 0.242∗∗∗ 0.483∗∗∗ 0.197∗∗∗ (0.018) (0.034) (0.049) (0.046) Deviance 16.729 58.186 120.155 106.944 Num. obs. 514 514 514 514 ∗∗∗p 0.001, ∗∗p 0.01, ∗p 0.05 Doesn’t look like parallel odds holds, maybe not ordered categories (or wrong categories)? 37 51
  • 39. Ordered Multinomial: Prediction Remember those cutoffs?! This is where they’re useful Assume Y has more than two ordered categories (for instance, low, medium, high) We now need two cut-points to divide the curve into three sections We’ll estimate these as µ1 and µ2 using MLE 38 51
  • 40. Ordered Multinomial: Prediction If Xiβ µ1 then predict Yi = low If µ1 Xiβ µ2 then predict Yi = medium If Xiβ µ2 then predict Yi = high β ∆ in x may change the prediction on Y, or it may not! 39 51
  • 41. Prediction in R: Creating data 1 # create fake data to use for prediction 2 predict _data − data . frame (mask = rep ( c ( No , Yes ) , 3) , 3 partyID = rep ( c ( An Independent , A Democrat , A Republican ) , each = 2) ) Wear mask party ID 1 No An Independent 2 Yes An Independent 3 No A Democrat 4 Yes A Democrat 5 No A Republican 6 Yes A Republican 40 51
  • 42. Prediction in R: Extract predicted values 1 # add in predictions for each observation in fake data 2 plot _data − melt ( cbind ( predict _data , predict ( ordered_ l o g i t _slim , predict _data , type = probs ) ) , id . vars = c ( mask , partyID ) , 3 variable . name = Level , value . name= Probability ) Wear mask party ID Level of concern Pr(Y = j) 1 No An Independent Not concerned at all 0.06 2 Yes An Independent Not concerned at all 0.02 3 No A Democrat Not concerned at all 0.03 4 Yes A Democrat Not concerned at all 0.01 . . . . . . . . . . . . 24 Yes A Republican Very concerned 0.27 41 51
  • 43. Prediction in R: Plot ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● Not concerned at all Not so concerned Somewhat concerned Very concerned A D e m o c r a t A R e p u b l i c a n A n I n d e p e n d e n t A D e m o c r a t A R e p u b l i c a n A n I n d e p e n d e n t A D e m o c r a t A R e p u b l i c a n A n I n d e p e n d e n t A D e m o c r a t A R e p u b l i c a n A n I n d e p e n d e n t 0.0 0.2 0.4 0.6 Party ID Pr(Y=j) Wear Mask ● ● No Yes 42 51
  • 44. Accuracy of predictions 1 # see how well our predictions were : 2 addmargins ( table ( covid_data$covid_concern , predict ( ordered_ l o g i t _slim , type= class ) ) ) Not concerned at all Not so concerned Somewhat concerned Very concerned Sum Not concerned at all 0 6 7 5 18 Not so concerned 0 18 38 21 77 Somewhat concerned 0 11 102 87 200 Very concerned 0 7 54 158 219 Sum 0 42 201 271 514 43 51
  • 45. Another Ex: Social Trust and Political Interest Is paying attention to politics associated with less trust in others? H1: The more people pay attention to negative political messages the less they trust others 1 # load in example 2 data : 2 # ESS p o l i t i c a l and social trust 3 # clean up data 4 # need to have registered email 5 set_email ( me@tcd. ie ) 6 # import a l l rounds from IE 7 ESS_data − import_ a l l _cntrounds ( country = Ireland ) 44 51
  • 46. Social Trust and Political Interest: Our Model Outcome: ppltrst: Most people can be trusted or you can’t be too careful I Values: Ordinal 0-11 Predictors: polintr: How interested in politics I Values: 1-4 (not at all, not really, somewhat, quite) essround: ESS round (time) 45 51
  • 47. Estimate model of trust and political interest 1 # run ordered l o g i t 2 ordered_ trust − polr ( ppltrst ~ polintr + essround , data = merged_ESS , Hess=T ) 3 4 # get odds ratios and CIs 5 exp ( cbind (OR = coef ( ordered_ trust ) , confint ( ordered_ trust ) ) ) OR 2.5 % 97.5 % polintr2 0.95 0.88 1.04 polintr3 0.76 0.69 0.83 polintr4 0.58 0.53 0.64 essround2 1.35 1.21 1.50 essround3 0.89 0.80 0.99 essround4 0.91 0.81 1.014 essround5 0.78 0.70 0.86 essround6 0.75 0.68 0.83 essround7 0.80 0.72 0.89 essround8 2.29 2.07 2.54 essround9 2.05 1.84 2.28 46 51
  • 48. Predicted probabilities with CIs 1 # get predicted probs for each category 2 # for person with high p o l i t i c a l interest ( 4 ) 3 # from last wave 4 # use l i b r a r y ( glm . predict to get CIs ) 5 basepredict . polr ( ordered_ trust , values = c ( rep (0 ,2) , 1 , rep ( 0 , 7 ) , 1 ) ) ¯ ŷ 2.5% 97.5% 0 0.021 0.018 0.023 1 0.027 0.025 0.030 2 0.040 0.037 0.044 3 0.064 0.059 0.069 4 0.077 0.071 0.082 5 0.151 0.144 0.158 6 0.144 0.139 0.149 7 0.176 0.169 0.182 8 0.180 0.169 0.191 9 0.076 0.070 0.083 10 0.035 0.032 0.039 11 0.009 0.007 0.011 47 51
  • 49. Change in Predicted probabilities with CIs 1 # see how predicted probs change between two cases 2 # low interest ( 1 ) to high ( 4 ) for wave 1 3 dc ( ordered_ trust , values1 = c ( rep (0 ,3) , 1 , rep ( 0 , 7 ) ) , 4 values2 = c ( rep (0 ,2) , 1 , 1 , rep ( 0 , 7 ) ) ) ¯ ∆ 2.5% 97.5% 0 -0.013 -0.015 -0.010 1 -0.016 -0.018 -0.013 2 -0.021 -0.024 -0.017 3 -0.028 -0.032 -0.023 4 -0.025 -0.029 -0.021 5 -0.028 -0.033 -0.022 6 -0.002 -0.005 0.001 7 0.026 0.021 0.030 8 0.055 0.046 0.064 9 0.031 0.025 0.037 10 0.016 0.013 0.019 11 0.004 0.003 0.005 48 51
  • 50. Wrap Up In this lesson, we went over how to... Extended logit regression for outcomes that have more than one category Extended multinomial logit regression for outcomes are ordered categories After reading week, we’ll talk about event count outcome variables 49 51
  • 51. Review: Logit regression estimation We want to think about how to take our set up from a binary logit and apply it to multiple outcome categories Without any loss of generality, index the Xi so that Xk is the variable and we’re still interested in taking exp of βk Fixing the values of each covariate and varying Xk by a small amount yields some δ log(p(..., Xk + δ)) − log(p(..., Xk)) = βkδ Thus, βk is marginal change in log odds with respect to Xk Back to slides 50 51
  • 52. Review: Logit regression estimation To recover exp(βk), we must set δ=1 and exponentiate left hand side: exp(βk) = exp(βk) × δ (1) = exp(βk) (2) = exp(log(p(..., Xk + δ)) − log(p(..., Xk))) (3) = p(..., Xk + 1) p(..., Xk) (4) (5) Back to slides 51 / 51