SlideShare una empresa de Scribd logo
1 de 24
Descargar para leer sin conexión
Habilitation `a diriger des recherches
Pierre Pudlo
Universit´e Montpellier 2
Institut de Math´ematiques et Mod´elisation de Montpellier (I3M)
Institut de Biologie Computationelle
Labex NUMEV
12/12/2014
Pierre Pudlo (UM2) HDR 12/12/2014 1 / 22
Contents
1 Graph based clustering
2 Approximate Bayesian computation
3 Bayesian computation with empirical likelihood
Pierre Pudlo (UM2) HDR 12/12/2014 2 / 22
Contents
1 Graph based clustering
2 Approximate Bayesian computation
3 Bayesian computation with empirical likelihood
Pierre Pudlo (UM2) HDR 12/12/2014 3 / 22
Graph based clustering
Idea. Define a (weighted) graph which links
similar observations Xi, i = 1, . . . , n.
Example: link Xi and Xj iff d(Xi, Xj) ≤ h
where h is a tuning parameter (bandwidth).
To obtain the clusters, cut the graph at a
minimal number of edges while partitioning
the data into large groups
The Ncut (Cheeger) for k = 2 groups.
Optimize
Ncut(A, ¯A) =
cut(A, ¯A)
min vol(A), vol( ¯A)
where ¯A is the complement of A in the set
of observations.
−2 −1 0 1 2
−1.0−0.50.00.51.0
Take into account vol(A) and vol( ¯A) to avoid
A = {a single outlier}
¯A = {the rest of the data set}
Pierre Pudlo (UM2) HDR 12/12/2014 4 / 22
A first asymptotic result
On spectral clustering.
The relaxed optimization problem
based on the spectral decomposition
of the matrix
Qh(i, j) =
kh(Xj − Xi)
kh(X − Xi)
• von Luxburg et al. (2008) obtained a
deterministic limit when
(∗) n → ∞ & h is fixed
Looks like the limit of kernel density
estimator fn,h(x), whose limit when h is
fixed is
f(y)kh(x − y)dy
• With Bruno Pelletier, we have
obtained asymptotic results under (∗).
We add a pre-processing: remove data
points in area of low density
Cluster the kept data points
Then, the limit does not depend on h &
can be characterized geometrically in
term of level sets of the true density
f(x).
• But the asymptotic behavior of
spectral clustering when
(∗∗) n → ∞ & h → 0
is much more difficult to
understand.
Pierre Pudlo (UM2) HDR 12/12/2014 5 / 22
A second asymptotic result
On the NP-hard optimization
problem which minimizes
Ncut(A, ¯A) =
cut(A, ¯A)
min vol(A), vol( ¯A)
on the graph with bandwidth h.
Assume M is a compact subset of Rd
,
X1, . . . Xn a uniform random sampling
of M.
(Known results if the points form a
regular grid of M)
• With Ery Arias-Castro and Bruno
Pelletier,
we have proven the asymptotic
consistency of the Ncut problem when
(∗ ∗ ∗) n → ∞ & h → 0,
h2d+1
1/
√
n
The limit is the solution of the Cheeger
problem
(cut M with a surface of minimal
perimeter)
Pierre Pudlo (UM2) HDR 12/12/2014 6 / 22
Contents
1 Graph based clustering
2 Approximate Bayesian computation
3 Bayesian computation with empirical likelihood
Pierre Pudlo (UM2) HDR 12/12/2014 7 / 22
Intractable likelihoods
Problem
How to perform a Bayesian analysis when the likelihood f(y|φ) is intractable?
Example 1. Gibbs random fields
f(y|φ) ∝ exp(−H(y, φ))
is known up to a constant
Z(φ) = 1
y
exp(−H(y, φ))
Example 2. Neutral population
genetics
Aim. Infer demographic parameters on
the past of some populations based on
the trace left in genomes of individuals
sampled from current populations.
Latent process (past history of the
sample) ∈ space of high dimension.
If y is the genetic data of the sample,
the likelihood is
f(y|φ) =
Z
f(y, z | φ) dz
Typically, dim(Z ) dim(Y ).
No hope to compute the likelihood with
clever Monte Carlo algorithms?
Pierre Pudlo (UM2) HDR 12/12/2014 8 / 22
Approximate Bayesian computation
Idea
Infer conditional distribution of φ given yobs on simulations from the joint π(φ)f(y|φ)
ABC algorithm
A) Generate a large set of (φ, y)
from the Bayesian model
π(φ) f(y|φ)
B) Keep the particles (φ, y) such
that d(η(yobs), η(y)) ≤ ε
C) Return the φ’s of the kept
particles
Curse of dimensionality: y is replaced
by some numerical summaries η(y)
Stage A) is computationally heavy!
We end up rejecting almost all
simulations except if fallen in the
neighborhood of η(yobs)
Sequential ABC algorithms try to avoid
drawing φ is area of low π(φ|y).
An auto-calibrated ABC-SMC
sampler with Mohammed Sedki,
Jean-Michel Marin, Jean-Marie Cornet
and Christian P. Robert
Pierre Pudlo (UM2) HDR 12/12/2014 9 / 22
ABC sequential sampler
How to calibrate ε1 ≥ ε2 ≥ · · · ≥ εT and T to be efficient?
The auto-calibrated ABC-SMC sampler developed with Mohammed Sedki,
Jean-Michel Marin, Jean-Marie Cornet and Christian P. Robert
Pierre Pudlo (UM2) HDR 12/12/2014 10 / 22
ABC target
Three levels of approximation of the
posterior π φ yobs
1 the ABC posterior distribution
π φ η(yobs)
2 approximated with a kernel of
bandwidth ε (or with k-nearest
neighbours)
π φ d(η(y), η(yobs)) ≤ ε
3 a Monte Carlo error:
sample size N < ∞
See, e.g., our review with J.-M. Marin,
C. Robert and R. Ryder
If η(y) are not sufficient statistics,
π φ yobs π φ η(yobs)
Information regarding yobs might be
lost!
Curse of dimensionality:
cannot have both ε small, N large
when η(y) is of large dimension
Post-processing of Beaumont et al.
(2002) with local linear regression.
But the lack of sufficiency might still be
problematic. See Robert et al. (2011)
for model choice.
Pierre Pudlo (UM2) HDR 12/12/2014 11 / 22
ABC model choice
ABC model choice
A) Generate a large set of
(m, φ, y) from the Bayesian
model, π(m)πm(φ) fm(y|φ)
B) Keep the particles (m, φ, y)
such that d(η(y), η(yobs)) ≤ ε
C) For each m, return
pm(yobs) = porportion of m
among the kept particles
Likewise, if η(y) is not sufficient for the
model choice issue,
π m y π m η(y)
It might be difficult to design
informative η(y).
Toy example.
Model 1. yi
iid
∼ N (φ, 1)
Model 2. yi
iid
∼ N (φ, 2)
Same prior on φ (whatever the model)
& uniform prior on model index
η(y) = y1 + · · · + yn is sufficient to
estimate φ in both models
But η(y) carries no information
regarding the variance (hence the
model choice issue)
Other examples in Robert et al. (2011)
In population genetics. Might be
difficult to find summary statistics that
help discriminate between models
(= possible historical scenarios on the
sampled populations)
Pierre Pudlo (UM2) HDR 12/12/2014 12 / 22
ABC model choice
ABC model choice
A) Generate a large set of
(m, φ, y) from the Bayesian
model π(m)πm(φ) fm(y|φ)
B) Keep the particles (m, φ, y)
such that d(η(y), η(yobs)) ≤ ε
C) For each m, return
pm(yobs) = porportion of m
among the kept particles
If ε is tuned so that the number of kept
particles is k, then pm is a k-nearest
neighbor estimate of
E 1{M = m} η(yobs)
Approximating the posterior
probabilities of model m is a
regression problem where
the response is 1{M = m},
the co-variables are the summary
statistics η(y),
the loss is L2
(conditional
expectation)
The prefered method to approximate
the postererior probabilities in DIYABC
is a local multinomial regression.
Ticklish if dim(η(y)) large, or high
correlation in the summary statistics.
Pierre Pudlo (UM2) HDR 12/12/2014 13 / 22
Choosing between hidden random fields
Choosing between dependency
graph: 4 or 8 neighbours?
Models. α, β ∼ prior
z | β ∼ Potts on G4 or G8 with interaction β
y | z, α ∼ i P(yi|zi, α)
How to sum up the noisy y?
Without noise (directly observed field),
sufficient statistics for the model choice
issue.
With Julien Stoehr and Lionel Cucala
a method to design new summary
statistics
Based on a clustering of the observed
data on possible dependency graphs
number of connected components
size of the largest connected
component,
. . .
Pierre Pudlo (UM2) HDR 12/12/2014 14 / 22
Machine learning to analyse machine simulated data
ABC model choice
A) Generate a large set of
(m, φ, y) from π(m)πm(φ) fm(y|φ)
B) Infer (anything?) about
m η(y) with machine learning
methods
In this machine learning perspective:
the (iid) simulations of A) form the
training set
yobs becomes a new data point
With J.-M. Marin, J.-M. Cornuet, A.
Estoup, M. Gautier and C. P. Robert
Predicting m is a classification
problem
Computing π(m|η(y)) is a
regression problem
It is well known that classification is
much simple than regression.
(dimension of the object we infer)
Why computing π(m|η(y)) if we know
that
π(m|y) π(m|η(y))?
Pierre Pudlo (UM2) HDR 12/12/2014 15 / 22
An example with random forest on human SNP data
Out of Africa
6 scenarios, 6 models
Observed data. 4 populations, 30
individuals per population; 10,000
genotyped SNP from the 1000 Genome
Project
Random forest trained on 40, 000
simulations (112 summary statistics)
predict the model which supports
a single out-of-Africa colonization
event,
a secondary split between European
and Asian lineages and
a recent admixture for Americans
with African origin
Confidence in the selected model?
Pierre Pudlo (UM2) HDR 12/12/2014 16 / 22
Example (continued)
Observed data. 4 populations, 30
individuals per population; 10,000
genotyped SNP from the 1000 Genome
Project
Random forest trained on 40, 000
simulations (112 summary statistics)
predict the model which supports
a single out-of-Africa colonization
event,
a secondary split between European
and Asian lineages and
a recent admixture for Americans
with African origin
Benefits of random forests?
1 Can find the relevant statistics in a
large set of statistics (112) to
discriminate models
2 Lower prior misclassification error
(≈ 6%) than other methods (ABC, i.e.
k-nn ≈ 18%)
3 Supply a similarity measure to
compare η(y) and η(yobs)
Confidence in the selected model?
Compute the average of the
misclassification error over an ABC
approximation of the predictive (∗). Here,
≤ 0.1%
(∗) π(m, φ, y | ηobs) = π(m | ηobs)πm(φ | ηobs)fm(y | φ)
Pierre Pudlo (UM2) HDR 12/12/2014 17 / 22
Contents
1 Graph based clustering
2 Approximate Bayesian computation
3 Bayesian computation with empirical likelihood
Pierre Pudlo (UM2) HDR 12/12/2014 18 / 22
Another approximation of the likelihood
What if both
the likelihood is intractable and
unable to simulate a dataset in a reasonable amount of time to resort on ABC?
First answer: use pseudo-likelihoods
such as the pairwise composite likelihood
fPCL(y | φ) =
i<j
f(yi, yj | φ)
Maximum composite likelihood
estimators φ(y) are suitable estimators
But cannot substitute a true likelihood
in a Bayesian framework
leads to credible intervals which are
too narrow: over-confidence in φ(y), see
e.g. Ribatet et al. (2012)
Our proposal with Kerrie Mengersen and
Christian P. Robert:
use the empirical likelihood of Owen
(2001, 2011)
It relies on iid blocks in the dataset y to
reconstruct a likelihood
& permits likelihood ratio tests
confidence intervals are correct
Original aim of Owen: remove parametric
assumptions
Pierre Pudlo (UM2) HDR 12/12/2014 19 / 22
Bayesian computation via empirical likelihood
Our proposal with Kerrie Mengersen and
Christian P. Robert:
use the empirical likelihood of Owen
(2001, 2011)
It relies on iid blocks in the dataset y to
reconstruct a likelihood
& permits likelihood ratio tests
confidence intervals are correct
Original aim of Owen: remove parametric
assumptions
With empirical likelihood, the parameter φ
is defined as
(∗) E h(yb, φ) = 0
where
yb is one block of y,
E the expected value according to
the true distribution of the block yb
h is a known function
E.g, if φ is the mean of an iid sample,
h(yb, φ) = yb − φ
In population genetics, what is (∗) with
dates of population splits
population sizes, etc. ?
Pierre Pudlo (UM2) HDR 12/12/2014 20 / 22
Bayesian computation via empirical likelihood
With empirical likelihood, the parameter φ
is defined as
(∗) E h(yb, φ) = 0
where
yb is one block of y,
E the expected value according to
the true distribution of the block yb
h is a known function
E.g, if φ is the mean of an iid sample,
h(yb, φ) = yb − φ
In population genetics, what is (∗) with
dates of population splits
population sizes, etc. ?
A block = genetic data at given locus
h(yb, φ) is the pairwise composite score
function we can explicitly compute in
many situations:
h(yb, φ) = φ log fPCL(yb | φ)
Benefits.
much faster than ABC (no need to
simulate fake data)
same accuracy than ABC or even
much precise: no loss of information
with summary statistics
Pierre Pudlo (UM2) HDR 12/12/2014 21 / 22
An experiment
Evolutionary scenario:
MRCA
POP 0 POP 1 POP 2
τ1
τ2
Dataset:
50 genes per populations,
100 microsat. loci
Assumptions:
Ne identical over all populations
φ = log10(θ, τ1, τ2)
non-informative prior
Comparison of ABC and EL
histogram = EL
curve = ABC
vertical line = “true” parameter
Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22
An experiment
Evolutionary scenario:
MRCA
POP 0 POP 1 POP 2
τ1
τ2
Dataset:
50 genes per populations,
100 microsat. loci
Assumptions:
Ne identical over all populations
φ = log10(θ, τ1, τ2)
non-informative prior
Comparison of ABC and EL
histogram = EL
curve = ABC
vertical line = “true” parameter
Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22
An experiment
Evolutionary scenario:
MRCA
POP 0 POP 1 POP 2
τ1
τ2
Dataset:
50 genes per populations,
100 microsat. loci
Assumptions:
Ne identical over all populations
φ = log10(θ, τ1, τ2)
non-informative prior
Comparison of ABC and EL
histogram = EL
curve = ABC
vertical line = “true” parameter
Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22

Más contenido relacionado

La actualidad más candente

Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
 
ABC short course: introduction chapters
ABC short course: introduction chaptersABC short course: introduction chapters
ABC short course: introduction chaptersChristian Robert
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015Christian Robert
 
Tailored Bregman Ball Trees for Effective Nearest Neighbors
Tailored Bregman Ball Trees for Effective Nearest NeighborsTailored Bregman Ball Trees for Effective Nearest Neighbors
Tailored Bregman Ball Trees for Effective Nearest NeighborsFrank Nielsen
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixturesChristian Robert
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationChristian Robert
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapterChristian Robert
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Frank Nielsen
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsChristian Robert
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testingChristian Robert
 
Slides: Hypothesis testing, information divergence and computational geometry
Slides: Hypothesis testing, information divergence and computational geometrySlides: Hypothesis testing, information divergence and computational geometry
Slides: Hypothesis testing, information divergence and computational geometryFrank Nielsen
 
Machine Learning in Actuarial Science & Insurance
Machine Learning in Actuarial Science & InsuranceMachine Learning in Actuarial Science & Insurance
Machine Learning in Actuarial Science & InsuranceArthur Charpentier
 
Stratified sampling and resampling for approximate Bayesian computation
Stratified sampling and resampling for approximate Bayesian computationStratified sampling and resampling for approximate Bayesian computation
Stratified sampling and resampling for approximate Bayesian computationUmberto Picchini
 
Approximating Bayes Factors
Approximating Bayes FactorsApproximating Bayes Factors
Approximating Bayes FactorsChristian Robert
 

La actualidad más candente (20)

Intractable likelihoods
Intractable likelihoodsIntractable likelihoods
Intractable likelihoods
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
ABC short course: introduction chapters
ABC short course: introduction chaptersABC short course: introduction chapters
ABC short course: introduction chapters
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
 
Side 2019 #12
Side 2019 #12Side 2019 #12
Side 2019 #12
 
Tailored Bregman Ball Trees for Effective Nearest Neighbors
Tailored Bregman Ball Trees for Effective Nearest NeighborsTailored Bregman Ball Trees for Effective Nearest Neighbors
Tailored Bregman Ball Trees for Effective Nearest Neighbors
 
SIAM CSE 2017 talk
SIAM CSE 2017 talkSIAM CSE 2017 talk
SIAM CSE 2017 talk
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimation
 
Georgia Tech 2017 March Talk
Georgia Tech 2017 March TalkGeorgia Tech 2017 March Talk
Georgia Tech 2017 March Talk
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapter
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
 
Slides: Hypothesis testing, information divergence and computational geometry
Slides: Hypothesis testing, information divergence and computational geometrySlides: Hypothesis testing, information divergence and computational geometry
Slides: Hypothesis testing, information divergence and computational geometry
 
Machine Learning in Actuarial Science & Insurance
Machine Learning in Actuarial Science & InsuranceMachine Learning in Actuarial Science & Insurance
Machine Learning in Actuarial Science & Insurance
 
Stratified sampling and resampling for approximate Bayesian computation
Stratified sampling and resampling for approximate Bayesian computationStratified sampling and resampling for approximate Bayesian computation
Stratified sampling and resampling for approximate Bayesian computation
 
MCQMC 2016 Tutorial
MCQMC 2016 TutorialMCQMC 2016 Tutorial
MCQMC 2016 Tutorial
 
Approximating Bayes Factors
Approximating Bayes FactorsApproximating Bayes Factors
Approximating Bayes Factors
 

Similar a Habilitation à diriger des recherches

Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesFrank Nielsen
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinChristian Robert
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep LearningRayKim51
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distancesChristian Robert
 
3rd NIPS Workshop on PROBABILISTIC PROGRAMMING
3rd NIPS Workshop on PROBABILISTIC PROGRAMMING3rd NIPS Workshop on PROBABILISTIC PROGRAMMING
3rd NIPS Workshop on PROBABILISTIC PROGRAMMINGChristian Robert
 
ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapterChristian Robert
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distancesChristian Robert
 
Pre-computation for ABC in image analysis
Pre-computation for ABC in image analysisPre-computation for ABC in image analysis
Pre-computation for ABC in image analysisMatt Moores
 
Approximate Bayesian computation for the Ising/Potts model
Approximate Bayesian computation for the Ising/Potts modelApproximate Bayesian computation for the Ising/Potts model
Approximate Bayesian computation for the Ising/Potts modelMatt Moores
 
Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image ijcsa
 
Iaetsd vlsi implementation of gabor filter based image edge detection
Iaetsd vlsi implementation of gabor filter based image edge detectionIaetsd vlsi implementation of gabor filter based image edge detection
Iaetsd vlsi implementation of gabor filter based image edge detectionIaetsd Iaetsd
 
Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionMichael Stumpf
 
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...NTNU
 

Similar a Habilitation à diriger des recherches (20)

Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective Divergences
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael Martin
 
QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
 
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
Deep Learning Opening Workshop - Horseshoe Regularization for Machine Learnin...
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
 
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
 
ABC based on Wasserstein distances
ABC based on Wasserstein distancesABC based on Wasserstein distances
ABC based on Wasserstein distances
 
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
 
3rd NIPS Workshop on PROBABILISTIC PROGRAMMING
3rd NIPS Workshop on PROBABILISTIC PROGRAMMING3rd NIPS Workshop on PROBABILISTIC PROGRAMMING
3rd NIPS Workshop on PROBABILISTIC PROGRAMMING
 
ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapter
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distances
 
Pre-computation for ABC in image analysis
Pre-computation for ABC in image analysisPre-computation for ABC in image analysis
Pre-computation for ABC in image analysis
 
asymptotics of ABC
asymptotics of ABCasymptotics of ABC
asymptotics of ABC
 
Approximate Bayesian computation for the Ising/Potts model
Approximate Bayesian computation for the Ising/Potts modelApproximate Bayesian computation for the Ising/Potts model
Approximate Bayesian computation for the Ising/Potts model
 
Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image Regularized Compression of A Noisy Blurred Image
Regularized Compression of A Noisy Blurred Image
 
Boston talk
Boston talkBoston talk
Boston talk
 
Iaetsd vlsi implementation of gabor filter based image edge detection
Iaetsd vlsi implementation of gabor filter based image edge detectionIaetsd vlsi implementation of gabor filter based image edge detection
Iaetsd vlsi implementation of gabor filter based image edge detection
 
Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model Selection
 
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
 

Último

Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...ssuser79fe74
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Lokesh Kothari
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformationAreesha Ahmad
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Bookingroncy bisnoi
 
FAIRSpectra - Enabling the FAIRification of Spectroscopy and Spectrometry
FAIRSpectra - Enabling the FAIRification of Spectroscopy and SpectrometryFAIRSpectra - Enabling the FAIRification of Spectroscopy and Spectrometry
FAIRSpectra - Enabling the FAIRification of Spectroscopy and SpectrometryAlex Henderson
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and ClassificationsAreesha Ahmad
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxCOST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxFarihaAbdulRasheed
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptxAlMamun560346
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...Lokesh Kothari
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxgindu3009
 
Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Silpa
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLkantirani197
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .Poonam Aher Patil
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 

Último (20)

Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
Chemical Tests; flame test, positive and negative ions test Edexcel Internati...
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
Labelling Requirements and Label Claims for Dietary Supplements and Recommend...
 
Conjugation, transduction and transformation
Conjugation, transduction and transformationConjugation, transduction and transformation
Conjugation, transduction and transformation
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Alandi Call Me 7737669865 Budget Friendly No Advance Booking
 
FAIRSpectra - Enabling the FAIRification of Spectroscopy and Spectrometry
FAIRSpectra - Enabling the FAIRification of Spectroscopy and SpectrometryFAIRSpectra - Enabling the FAIRification of Spectroscopy and Spectrometry
FAIRSpectra - Enabling the FAIRification of Spectroscopy and Spectrometry
 
Site Acceptance Test .
Site Acceptance Test                    .Site Acceptance Test                    .
Site Acceptance Test .
 
Bacterial Identification and Classifications
Bacterial Identification and ClassificationsBacterial Identification and Classifications
Bacterial Identification and Classifications
 
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptxCOST ESTIMATION FOR A RESEARCH PROJECT.pptx
COST ESTIMATION FOR A RESEARCH PROJECT.pptx
 
Seismic Method Estimate velocity from seismic data.pptx
Seismic Method Estimate velocity from seismic  data.pptxSeismic Method Estimate velocity from seismic  data.pptx
Seismic Method Estimate velocity from seismic data.pptx
 
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
GUIDELINES ON SIMILAR BIOLOGICS Regulatory Requirements for Marketing Authori...
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptx
 
Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.Proteomics: types, protein profiling steps etc.
Proteomics: types, protein profiling steps etc.
 
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRLKochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
Kochi ❤CALL GIRL 84099*07087 ❤CALL GIRLS IN Kochi ESCORT SERVICE❤CALL GIRL
 
Factory Acceptance Test( FAT).pptx .
Factory Acceptance Test( FAT).pptx       .Factory Acceptance Test( FAT).pptx       .
Factory Acceptance Test( FAT).pptx .
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 

Habilitation à diriger des recherches

  • 1. Habilitation `a diriger des recherches Pierre Pudlo Universit´e Montpellier 2 Institut de Math´ematiques et Mod´elisation de Montpellier (I3M) Institut de Biologie Computationelle Labex NUMEV 12/12/2014 Pierre Pudlo (UM2) HDR 12/12/2014 1 / 22
  • 2. Contents 1 Graph based clustering 2 Approximate Bayesian computation 3 Bayesian computation with empirical likelihood Pierre Pudlo (UM2) HDR 12/12/2014 2 / 22
  • 3. Contents 1 Graph based clustering 2 Approximate Bayesian computation 3 Bayesian computation with empirical likelihood Pierre Pudlo (UM2) HDR 12/12/2014 3 / 22
  • 4. Graph based clustering Idea. Define a (weighted) graph which links similar observations Xi, i = 1, . . . , n. Example: link Xi and Xj iff d(Xi, Xj) ≤ h where h is a tuning parameter (bandwidth). To obtain the clusters, cut the graph at a minimal number of edges while partitioning the data into large groups The Ncut (Cheeger) for k = 2 groups. Optimize Ncut(A, ¯A) = cut(A, ¯A) min vol(A), vol( ¯A) where ¯A is the complement of A in the set of observations. −2 −1 0 1 2 −1.0−0.50.00.51.0 Take into account vol(A) and vol( ¯A) to avoid A = {a single outlier} ¯A = {the rest of the data set} Pierre Pudlo (UM2) HDR 12/12/2014 4 / 22
  • 5. A first asymptotic result On spectral clustering. The relaxed optimization problem based on the spectral decomposition of the matrix Qh(i, j) = kh(Xj − Xi) kh(X − Xi) • von Luxburg et al. (2008) obtained a deterministic limit when (∗) n → ∞ & h is fixed Looks like the limit of kernel density estimator fn,h(x), whose limit when h is fixed is f(y)kh(x − y)dy • With Bruno Pelletier, we have obtained asymptotic results under (∗). We add a pre-processing: remove data points in area of low density Cluster the kept data points Then, the limit does not depend on h & can be characterized geometrically in term of level sets of the true density f(x). • But the asymptotic behavior of spectral clustering when (∗∗) n → ∞ & h → 0 is much more difficult to understand. Pierre Pudlo (UM2) HDR 12/12/2014 5 / 22
  • 6. A second asymptotic result On the NP-hard optimization problem which minimizes Ncut(A, ¯A) = cut(A, ¯A) min vol(A), vol( ¯A) on the graph with bandwidth h. Assume M is a compact subset of Rd , X1, . . . Xn a uniform random sampling of M. (Known results if the points form a regular grid of M) • With Ery Arias-Castro and Bruno Pelletier, we have proven the asymptotic consistency of the Ncut problem when (∗ ∗ ∗) n → ∞ & h → 0, h2d+1 1/ √ n The limit is the solution of the Cheeger problem (cut M with a surface of minimal perimeter) Pierre Pudlo (UM2) HDR 12/12/2014 6 / 22
  • 7. Contents 1 Graph based clustering 2 Approximate Bayesian computation 3 Bayesian computation with empirical likelihood Pierre Pudlo (UM2) HDR 12/12/2014 7 / 22
  • 8. Intractable likelihoods Problem How to perform a Bayesian analysis when the likelihood f(y|φ) is intractable? Example 1. Gibbs random fields f(y|φ) ∝ exp(−H(y, φ)) is known up to a constant Z(φ) = 1 y exp(−H(y, φ)) Example 2. Neutral population genetics Aim. Infer demographic parameters on the past of some populations based on the trace left in genomes of individuals sampled from current populations. Latent process (past history of the sample) ∈ space of high dimension. If y is the genetic data of the sample, the likelihood is f(y|φ) = Z f(y, z | φ) dz Typically, dim(Z ) dim(Y ). No hope to compute the likelihood with clever Monte Carlo algorithms? Pierre Pudlo (UM2) HDR 12/12/2014 8 / 22
  • 9. Approximate Bayesian computation Idea Infer conditional distribution of φ given yobs on simulations from the joint π(φ)f(y|φ) ABC algorithm A) Generate a large set of (φ, y) from the Bayesian model π(φ) f(y|φ) B) Keep the particles (φ, y) such that d(η(yobs), η(y)) ≤ ε C) Return the φ’s of the kept particles Curse of dimensionality: y is replaced by some numerical summaries η(y) Stage A) is computationally heavy! We end up rejecting almost all simulations except if fallen in the neighborhood of η(yobs) Sequential ABC algorithms try to avoid drawing φ is area of low π(φ|y). An auto-calibrated ABC-SMC sampler with Mohammed Sedki, Jean-Michel Marin, Jean-Marie Cornet and Christian P. Robert Pierre Pudlo (UM2) HDR 12/12/2014 9 / 22
  • 10. ABC sequential sampler How to calibrate ε1 ≥ ε2 ≥ · · · ≥ εT and T to be efficient? The auto-calibrated ABC-SMC sampler developed with Mohammed Sedki, Jean-Michel Marin, Jean-Marie Cornet and Christian P. Robert Pierre Pudlo (UM2) HDR 12/12/2014 10 / 22
  • 11. ABC target Three levels of approximation of the posterior π φ yobs 1 the ABC posterior distribution π φ η(yobs) 2 approximated with a kernel of bandwidth ε (or with k-nearest neighbours) π φ d(η(y), η(yobs)) ≤ ε 3 a Monte Carlo error: sample size N < ∞ See, e.g., our review with J.-M. Marin, C. Robert and R. Ryder If η(y) are not sufficient statistics, π φ yobs π φ η(yobs) Information regarding yobs might be lost! Curse of dimensionality: cannot have both ε small, N large when η(y) is of large dimension Post-processing of Beaumont et al. (2002) with local linear regression. But the lack of sufficiency might still be problematic. See Robert et al. (2011) for model choice. Pierre Pudlo (UM2) HDR 12/12/2014 11 / 22
  • 12. ABC model choice ABC model choice A) Generate a large set of (m, φ, y) from the Bayesian model, π(m)πm(φ) fm(y|φ) B) Keep the particles (m, φ, y) such that d(η(y), η(yobs)) ≤ ε C) For each m, return pm(yobs) = porportion of m among the kept particles Likewise, if η(y) is not sufficient for the model choice issue, π m y π m η(y) It might be difficult to design informative η(y). Toy example. Model 1. yi iid ∼ N (φ, 1) Model 2. yi iid ∼ N (φ, 2) Same prior on φ (whatever the model) & uniform prior on model index η(y) = y1 + · · · + yn is sufficient to estimate φ in both models But η(y) carries no information regarding the variance (hence the model choice issue) Other examples in Robert et al. (2011) In population genetics. Might be difficult to find summary statistics that help discriminate between models (= possible historical scenarios on the sampled populations) Pierre Pudlo (UM2) HDR 12/12/2014 12 / 22
  • 13. ABC model choice ABC model choice A) Generate a large set of (m, φ, y) from the Bayesian model π(m)πm(φ) fm(y|φ) B) Keep the particles (m, φ, y) such that d(η(y), η(yobs)) ≤ ε C) For each m, return pm(yobs) = porportion of m among the kept particles If ε is tuned so that the number of kept particles is k, then pm is a k-nearest neighbor estimate of E 1{M = m} η(yobs) Approximating the posterior probabilities of model m is a regression problem where the response is 1{M = m}, the co-variables are the summary statistics η(y), the loss is L2 (conditional expectation) The prefered method to approximate the postererior probabilities in DIYABC is a local multinomial regression. Ticklish if dim(η(y)) large, or high correlation in the summary statistics. Pierre Pudlo (UM2) HDR 12/12/2014 13 / 22
  • 14. Choosing between hidden random fields Choosing between dependency graph: 4 or 8 neighbours? Models. α, β ∼ prior z | β ∼ Potts on G4 or G8 with interaction β y | z, α ∼ i P(yi|zi, α) How to sum up the noisy y? Without noise (directly observed field), sufficient statistics for the model choice issue. With Julien Stoehr and Lionel Cucala a method to design new summary statistics Based on a clustering of the observed data on possible dependency graphs number of connected components size of the largest connected component, . . . Pierre Pudlo (UM2) HDR 12/12/2014 14 / 22
  • 15. Machine learning to analyse machine simulated data ABC model choice A) Generate a large set of (m, φ, y) from π(m)πm(φ) fm(y|φ) B) Infer (anything?) about m η(y) with machine learning methods In this machine learning perspective: the (iid) simulations of A) form the training set yobs becomes a new data point With J.-M. Marin, J.-M. Cornuet, A. Estoup, M. Gautier and C. P. Robert Predicting m is a classification problem Computing π(m|η(y)) is a regression problem It is well known that classification is much simple than regression. (dimension of the object we infer) Why computing π(m|η(y)) if we know that π(m|y) π(m|η(y))? Pierre Pudlo (UM2) HDR 12/12/2014 15 / 22
  • 16. An example with random forest on human SNP data Out of Africa 6 scenarios, 6 models Observed data. 4 populations, 30 individuals per population; 10,000 genotyped SNP from the 1000 Genome Project Random forest trained on 40, 000 simulations (112 summary statistics) predict the model which supports a single out-of-Africa colonization event, a secondary split between European and Asian lineages and a recent admixture for Americans with African origin Confidence in the selected model? Pierre Pudlo (UM2) HDR 12/12/2014 16 / 22
  • 17. Example (continued) Observed data. 4 populations, 30 individuals per population; 10,000 genotyped SNP from the 1000 Genome Project Random forest trained on 40, 000 simulations (112 summary statistics) predict the model which supports a single out-of-Africa colonization event, a secondary split between European and Asian lineages and a recent admixture for Americans with African origin Benefits of random forests? 1 Can find the relevant statistics in a large set of statistics (112) to discriminate models 2 Lower prior misclassification error (≈ 6%) than other methods (ABC, i.e. k-nn ≈ 18%) 3 Supply a similarity measure to compare η(y) and η(yobs) Confidence in the selected model? Compute the average of the misclassification error over an ABC approximation of the predictive (∗). Here, ≤ 0.1% (∗) π(m, φ, y | ηobs) = π(m | ηobs)πm(φ | ηobs)fm(y | φ) Pierre Pudlo (UM2) HDR 12/12/2014 17 / 22
  • 18. Contents 1 Graph based clustering 2 Approximate Bayesian computation 3 Bayesian computation with empirical likelihood Pierre Pudlo (UM2) HDR 12/12/2014 18 / 22
  • 19. Another approximation of the likelihood What if both the likelihood is intractable and unable to simulate a dataset in a reasonable amount of time to resort on ABC? First answer: use pseudo-likelihoods such as the pairwise composite likelihood fPCL(y | φ) = i<j f(yi, yj | φ) Maximum composite likelihood estimators φ(y) are suitable estimators But cannot substitute a true likelihood in a Bayesian framework leads to credible intervals which are too narrow: over-confidence in φ(y), see e.g. Ribatet et al. (2012) Our proposal with Kerrie Mengersen and Christian P. Robert: use the empirical likelihood of Owen (2001, 2011) It relies on iid blocks in the dataset y to reconstruct a likelihood & permits likelihood ratio tests confidence intervals are correct Original aim of Owen: remove parametric assumptions Pierre Pudlo (UM2) HDR 12/12/2014 19 / 22
  • 20. Bayesian computation via empirical likelihood Our proposal with Kerrie Mengersen and Christian P. Robert: use the empirical likelihood of Owen (2001, 2011) It relies on iid blocks in the dataset y to reconstruct a likelihood & permits likelihood ratio tests confidence intervals are correct Original aim of Owen: remove parametric assumptions With empirical likelihood, the parameter φ is defined as (∗) E h(yb, φ) = 0 where yb is one block of y, E the expected value according to the true distribution of the block yb h is a known function E.g, if φ is the mean of an iid sample, h(yb, φ) = yb − φ In population genetics, what is (∗) with dates of population splits population sizes, etc. ? Pierre Pudlo (UM2) HDR 12/12/2014 20 / 22
  • 21. Bayesian computation via empirical likelihood With empirical likelihood, the parameter φ is defined as (∗) E h(yb, φ) = 0 where yb is one block of y, E the expected value according to the true distribution of the block yb h is a known function E.g, if φ is the mean of an iid sample, h(yb, φ) = yb − φ In population genetics, what is (∗) with dates of population splits population sizes, etc. ? A block = genetic data at given locus h(yb, φ) is the pairwise composite score function we can explicitly compute in many situations: h(yb, φ) = φ log fPCL(yb | φ) Benefits. much faster than ABC (no need to simulate fake data) same accuracy than ABC or even much precise: no loss of information with summary statistics Pierre Pudlo (UM2) HDR 12/12/2014 21 / 22
  • 22. An experiment Evolutionary scenario: MRCA POP 0 POP 1 POP 2 τ1 τ2 Dataset: 50 genes per populations, 100 microsat. loci Assumptions: Ne identical over all populations φ = log10(θ, τ1, τ2) non-informative prior Comparison of ABC and EL histogram = EL curve = ABC vertical line = “true” parameter Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22
  • 23. An experiment Evolutionary scenario: MRCA POP 0 POP 1 POP 2 τ1 τ2 Dataset: 50 genes per populations, 100 microsat. loci Assumptions: Ne identical over all populations φ = log10(θ, τ1, τ2) non-informative prior Comparison of ABC and EL histogram = EL curve = ABC vertical line = “true” parameter Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22
  • 24. An experiment Evolutionary scenario: MRCA POP 0 POP 1 POP 2 τ1 τ2 Dataset: 50 genes per populations, 100 microsat. loci Assumptions: Ne identical over all populations φ = log10(θ, τ1, τ2) non-informative prior Comparison of ABC and EL histogram = EL curve = ABC vertical line = “true” parameter Pierre Pudlo (UM2) HDR 12/12/2014 22 / 22