Tutorial: Evaluation of Galaxy Cluster Centering Models¶
Previously, we have mostly done simple, visual comparisons of fitted models or posterior predictions with data. In this notebook, we'll get more quantitative about saying whether a model could plausibly produce the observed data. Specifically, for a simple data set encoding the difference between 2 methods of defining a galaxy cluster's center, you will
- fit a simple model, visually compare the posterior to the data, and quantitatively compare posterior predictions of a test statistic with the data;
- propose a more complex model and repeat the same comparisons;
- compare the 2 models using the Deviance Information Criterion;
- compare the 2 models using the Bayesian evidence
# !pip install emcee incredible pygtc
from os import getcwd
from os.path import exists as file_exists
from yaml import safe_load
import numpy as np
from scipy.special import logsumexp
import scipy.stats as st
import matplotlib.pyplot as plt
%matplotlib inline
import emcee
import incredible as cr
from pygtc import plotGTC
thisTutorial = 'model_evaluation'
if getcwd() == '/content':
# assume we are in Colab, and the user's data directory is linked to their drive/Physics267_data
from google.colab import drive
drive.mount('/content/drive')
datapath = '/content/drive/MyDrive/Physics267_data/' + thisTutorial + '/'
else:
# assume we are running locally somewhere and have the data under ./data/
datapath = 'data/'
The Dataset¶
Our data is just a list of numbers. Each one represents a measured distance, $y$, between two different estimates of the center of a galaxy cluster: the location of the presumed central galaxy and a centroid of the diffuse, emissive gas. The context here is that automated algorithms sometimes fail to chose the central galaxy correctly (because of image artifacts or other problems), whereas the gas centroid is more reliable but also more expensive to measure. Therefore, we'd like to use this data set to characterize the distribution of mis-centerings so that the galaxy-based centers can be used for large sample, with the resulting errors propagated forward through future processing, e.g. weak lensing estimates.
Let's load up the data and have a look. Each entry in y
is an offset in units of kpc. As usual, your particular data set is fictional, but intentionally similar to a real one that has actually been used for this purpose.
y = np.loadtxt(datapath+'offsets.txt')
Check out a quick histogram of the data.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
bins = np.linspace(0,1000,20)
plt.hist(y, bins=bins);
plt.xlabel("Offset $y$ (kpc)", fontsize=12);
plt.ylabel("Frequency", fontsize=12);
1. Choosing a Test Statistic¶
The model we will test in this tutorial is outlined in the next section - but how well that model fits the data is a question we will answer in part using a test statistic.
In this case, a feature we especially want to capture with our model is the fraction of clusters where $y$ is small (in the context of a galaxy cluster) vs not small. For concreteness, in the work you turn in, make $T$ the number of clusters with $y<100$ kpc. However, you are encouraged to play around with other possibilities to see how the simple model below succeeds or fails to reproduce the corresponding features of the data.
Below, define a function that computes this test statistic.
def T(yy):
"""
Argument: a data vector (either the real data or a simulated data set)
Returns: a scalar test statistic computed from the argument
"""
# YOUR CODE HERE
raise NotImplementedError()
Compute the test statistic of the real data to verify that it works.
T_true = T(y)
print("The test statistic of the real data is T =", T_true)
Setting up a Computational Framework¶
Once we define a model to work with (below), we'll want to fit that model to the data, and then evaluate it using the methods we saw in the Model Evaluation notes. These include:
- a visual check using replica datasets drawn from the posterior predictive distribution
- a quantitative posterior predictive model check using a suitable test statistic $T(y)$
In the next notebook, we'll choose and fit a second, alternative model, and compare the two in terms of
- the Deviance Information Criterion (DIC), to assess the models' (relative) generalized predictive accuracy
- the Bayesian Evidence, to provide insight on the (relative) probability of each model given the data
Notice that each of these bulleted operations can be coded as a function of the model, $H$ (e.g. a visual check of the model, the evidence of the model, and so on). We will therefore be fancy and write a class that completely describes the model, and also contains a set of functions that act on the model (methods). Since we anticipate looking at multiple models, we'll use inheritance. This allows us to automatically reuse methods that don't depend on the specific model in question. While this level of object oriented programming may not be familiar, it's a good thing to learn, and most of the details are filled in for you below.
We start by defining a base class, which contains the functionality common to any model we care to define later. To make it clear what functionality we expect derived classes to provide, we'll include defintions of non-functional methods that the derived classes will need to override. (This is completely unnecessary in Python, but has the benefit of providing a complete list of functions we'll eventually want to fill in.)
The functions that are filled in below (towards the bottom) are the ones that we won't need to modify for any particular model. Do take a moment to make sure you understand how they work; they are mostly doing things you've seen before. Wherever you see **params
here, a derived class would have a specific list of arguments corresponding to the model parameters.
# This is something we can throw to discourage direct instantiation of the base class
class VirtualClassError(Exception):
def __init__(self):
Exception.__init__(self,"Do not directly instantiate the base Model class!")
class Model:
"""
Base class for inference and model evaluation in a simple cluster mis-centering analysis.
In all these functions, `params' is a dictionary of model parameters.
"""
def __init__(self, priors, samples=None):
# Storage for MCMC samples from fitting the model
self.samples = samples
if samples is None:
self.Nsamples = 0
else:
self.Nsamples = samples.shape[0]
# a dictionary of priors (keys match param_names, values should have scipy distribution-like functionality)
# for simplicity, assume independent priors for each parameter
self.priors = priors
# The __init__ for derived classes should not repeat the lines above, but should provide what follows:
# # A name for the model!
# self.name =
# # As part of the model definition, store an ordered list of parameter names (for dictionary keys)
# # Named arguments of other functions that correspond to parameters should appear in this order
# self.param_names =
# # parameter names as display text (for plot labels),
# self.param_labels =
# # The next line finishes initialization by calling the parent class' __init__
# Model.__init__(self, *args, **kwargs)
def log_likelihood(self, **params):
"""
Return the log of the likelihood function L(params) = p(y|params,H)
"""
raise VirtualClassError # to be overriden by child classes
def sampling_distribution(self, yy, **params):
"""
Return the sampling distribution p(yy|params,H) at a point in data space yy given parameter(s) args
We expect a vector input yy, and return the corresponding probabilities.
Note: This is useful for making plots of "the model" overlaid on the histogram of the data
"""
raise VirtualClassError # to be overriden by child classes
def generate_replica_dataset(self, N, **params):
"""
Return a replica dataset y_rep of length N from the sampling distribution p(y_rep|args,H) as an array.
"""
raise VirtualClassError # to be overriden by child classes
def log_prior(self, **params):
"""
Return the log prior PDF p(params|H)
"""
return np.sum([self.priors[k].logpdf(params[k]) for k in params.keys()])
def draw_samples_from_prior(self, N):
"""
Return N samples from the prior PDF p(params|H) as a list of dictionaries
"""
return [{k:self.priors[k].rvs() for k in self.param_names} for i in range(N)]
def log_posterior(self, parameterlist=None, **params):
"""
Return the log of the (unnormalized) posterior PDF p(params|y,H)
The parameterlist argument is there for compatibility with emcee.
This will be an infinite loop if you named one of your parameters "parameterlist", but in that case
you deserve it.
This also means that for even a simple call, we would need to specify the parameter name(s),
i.e. not log_posterior(pval), but log_posterior(param=pval). This doesn't seem unreasonable.
"""
if parameterlist is not None:
pdict = {k:parameterlist[i] for i,k in enumerate(self.param_names)}
return self.log_posterior(**pdict)
lnp = self.log_prior(**params)
if lnp != -np.inf:
lnp += self.log_likelihood(**params)
return lnp
def draw_samples_from_posterior(self, starting_params, nsteps, nwalkers=8, threads=1):
"""
Use emcee to draw samples from P(params|y,H). Hopefully it Just Works.
You could try using e.g. threads=4 to speed things up with multiprocessing.
"""
# The density to sample is this model's own posterior PDF
npars = len(starting_params)
self.sampler = emcee.EnsembleSampler(nwalkers, npars, self.log_posterior, threads=threads)
# Generate an ensemble of walkers within +/-1% of the guess:
theta_0 = np.array([starting_params*(1.0 + 0.01*np.random.randn(npars)) for j in range(nwalkers)])
# Note that the initial parameter array theta_0 should have dimensions nwalkers x npars
# Evolve the ensemble:
self.sampler.run_mcmc(theta_0, nsteps)
# Plot the raw samples:
plt.rcParams['figure.figsize'] = (12.0, 4.0*npars)
fig, ax = plt.subplots(npars, 1);
cr.plot_traces(self.sampler.chain[:min(8,nwalkers),:,:], ax, labels=self.param_labels);
def check_chains(self, burn, maxlag):
'''
Ignoring `burn` samples from the front of each chain, compute convergence criteria and
effective number of samples.
'''
nwalk, nsteps, npars = self.sampler.chain.shape
if burn < 1 or burn >= nsteps:
return
tmp_samples = [self.sampler.chain[i,burn:,:] for i in range(nwalk)]
self.R = cr.GelmanRubinR(tmp_samples)
self.neff = cr.effective_samples(tmp_samples, maxlag=maxlag)
print('R =', self.R)
print('neff =', self.neff)
print('NB: Since walkers are not independent, these will be optimistic!')
def remove_burnin(self, burn):
'''
Remove `burn` samples from the front of each chain, and concatenate
Plot, and store the result in self.samples
'''
nwalk, nsteps, npars = self.sampler.chain.shape
if burn < 1 or burn >= nsteps:
return
self.samples = self.sampler.chain[:,burn:,:].reshape(nwalk*(nsteps-burn), npars)
self.Nsamples = self.samples.shape[0]
plt.rcParams['figure.figsize'] = (12.0, 4.0*npars)
fig, ax = plt.subplots(npars, 1);
cr.plot_traces(self.samples, ax, labels=self.param_labels);
def posterior_mean(self):
'''
Helper function for computing the posterior mean of each parameter (from MCMC samples)
'''
m = np.mean(self.samples, axis=0)
return {k:m[i] for i,k in enumerate(self.param_names)}
2. Evaluating a Simple Model¶
First, let's assume the simple model $H_1$, that the sampling distribution is an exponential:
$p(\{y_i\}|a, H_1) = \prod_i \frac{1}{a}e^{-y_i/a}$; $y_i\geq0$
Our single parameter is $a$, the mean of the exponential distribution.
Note that this is the entirety of the sampling distribution. Don't be scared of the simplicity! Our model is that the data are realizations of some distribution, and we want (initially) to find the parameter(s) of that distribution. There's nothing tricky (but also nothing Gaussian) about this.
As usual, draw the PGM and specify all the probabilistic relationships in the model. Note: In class we will discuss and agree on a common set of priors that everyone should use.
Space for PGM and generative model
2a. Implementation in code¶
We provide an implementation of the ExponentialModel
class, derived from Model
, below. As advertized, only the methods that depend on the sampling distribution need to be specifically implemented; the other Model
methods are still available. You will still need to specify the priors for this model when we instantiate it.
class ExponentialModel(Model):
"""
Simple exponential model for mis-centering.
"""
def __init__(self, *args, **kwargs):
# A name for the model!
self.name = 'Exponential Model'
# Named arguments of other functions that correspond to parameters should appear in this order
self.param_names = ['a']
# parameter names as display text (for plot labels),
self.param_labels = [r'$a$']
# The next line finishes initialization by calling the parent class' __init__
Model.__init__(self, *args, **kwargs)
def log_likelihood(self, a):
"""
Evaluate the log of the likelihood function L(a) = p(y|a,H)
"""
return np.sum(st.expon.logpdf(y, scale=a))
def sampling_distribution(self, yy, a):
"""
Evaluate the sampling distribution PDF p(yy|a,H) at a point in data space yy given parameter value a
We expect a vector input yy, and return the corresponding probabilities.
Note: This is useful for making plots of "the model" overlaid on the histogram of the data
"""
return st.expon.pdf(yy, scale=a)
def generate_replica_dataset(self, N, a):
"""
Return a replica data set y_rep of length N from the sampling distribution p(y_rep|a,H).
"""
return st.expon.rvs(scale=a, size=N)
Per the comments in Model.__init__
, define a dictionary whose sole entry, expon_priors['a']
, specifies the prior distribution for $a$. This will need to be in the form of a scipy.stats
frozen distribution, or some other object with many of the same methods.
# expon_priors = {'a':...}
# YOUR CODE HERE
raise NotImplementedError()
Now try instantiating a model object:
Model1 = ExponentialModel(expon_priors)
Try drawing a dozen samples from its prior as a test:
print("12 sample values drawn from the prior of Model1: ", Model1.draw_samples_from_prior(12))
Test out the log-posterior function to make sure it's not obviously buggy:
for a in [1.0, 10.0, 100.0, -3.14]:
print('Log-posterior for a=', a, ' = ', Model1.log_posterior(a=a))
Similarly the mock-data producing function (arbitrarily with $a=500$).
plt.rcParams['figure.figsize'] = (6.0, 4.0)
plt.hist(Model1.generate_replica_dataset(N=len(y), a=500.), bins=bins, color="lightgray");
plt.xlabel("Center offset $y$ (kpc)", fontsize=12);
plt.ylabel("Frequency (replica)", fontsize=12);
Finally, test the sampling distribution function.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
plt.plot(bins, Model1.sampling_distribution(bins, a=500.));
plt.xlabel("Center offset $y$ (kpc)", fontsize=12);
plt.ylabel("$p(y|a)$", fontsize=12);
2b. Fit the model to the data¶
The draw_samples_from_posterior
method carries out a parameter inference with emcee
and displays its Markov chains. remove_burnin
removes burn-in and concatenates the chains. Since these step aren't really the point of this problem, we'll just give you the code... for this model. Keep in mind that you're going to choose the alternative model to consider below, so we can't say what values of the keyword arguments used below will be appropriate for that case. Best not to speed through this section without looking at and thinking about what's happening, in other words.
The concatenated MCMC samples are eventually stored in the Model1.samples
array. Since we will be using these to compute p-values, among other things, make sure there are at least $10^3$ effectively independent samples after burn-in.
%%time
Model1.draw_samples_from_posterior(starting_params=[100.0], nsteps=2500, nwalkers=16)
This will compute the usual convergence criteria and effective number of samples:
Model1.check_chains(burn=200, maxlag=200)
Assuming that looks ok, remove the burn-in for good, and plot the concatenation of what's left.
Model1.remove_burnin(burn=100)
assert Model1.neff[0] >= 1000
It will be useful for later to know the mean of the posterior:
print("Posterior mean:", Model1.posterior_mean())
This_all_looks_good = False # change to True when true
# YOUR CODE HERE
raise NotImplementedError()
assert This_all_looks_good
2c. Visually compare the posterior predictions with the data.¶
First, let's just plot the posterior-mean model over the data.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
# First the histogram of observed data, as backdrop:
plt.hist(y, bins=bins, color="skyblue", density=True, label="Observed")
# Now overlay a curve following the sampling distribution conditioned on the posterior mean value of a1:
yy = np.linspace(bins.min(), bins.max(), len(bins)*10)
pp = Model1.sampling_distribution(bins, **Model1.posterior_mean())
plt.plot(bins, pp, linestyle="dashed", color="red", label="Posterior mean model")
plt.xlabel("Offset $y$ (kpc)", fontsize=12)
plt.ylabel("Normalized Frequency", fontsize=12)
plt.legend();
This kind of plot should be familiar: it's often a good idea to evaluate model adequacy in data space. You should already be able to see telling differences between the a well-fitting model's sampling distribution, and the data histogram. However, we don't know to what extent those could be explained simply by the variability in the sampling distribution, let along the posterior uncertainty in $a$.
So, next, let's compare a random predicted ("replica") data set, drawn from the posterior predictive distribution, with the data. To do this we first draw a parameter value from the posterior PDF, and then generate a dataset from the sampling distribution conditioned on that value. The result is a sample from $p(y_\mathrm{rep}|y,H_1)$.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
# First the histogram of observed data, as backdrop:
plt.hist(y, bins=bins, color="skyblue", density=True, label="Observed");
# Choose a posterior sample at random and generate a replica dataset, and show its histogram
j = np.random.randint(0, Model1.Nsamples)
mock = Model1.generate_replica_dataset(len(y), *Model1.samples[j,:])
plt.hist(mock, bins=bins, alpha=1.0, histtype="step", color="red", density=True, label="Sample posterior prediction")
plt.xlabel("Offset $y$ (kpc)", fontsize=12)
plt.ylabel("Normalized Frequency", fontsize=12)
plt.legend();
This plot is nice because it is comparing apples with apples, asking: do mock datasets drawn from our model sampling distribution with any plausible parameter value "look like" the real data?
However, to best evaluate this, we want to visualize the posterior predictive distribution of replica datasets instead of looking at just one. We can do this by plotting many replica datasets drawn from the posterior predictive PDF. Let's put this plot in a function, so we can re-use it later.
def visual_check(axes, Model, Nreps=None):
# First the histogram of observed data, as backdrop:
axes.hist(y, bins=bins, color="skyblue", density=True, label="Observed");
# Compute the posterior mean parameters
pm = Model.posterior_mean()
# Make a large number of replica datasets, and overlay histograms of them all
if Nreps is None: Nreps = len(Model.samples)
alpha = 5.0 / Nreps
for jj in np.round(np.linspace(0, len(Model.samples), num=Nreps, endpoint=False)):
j = int(jj)
if j==0:
# Plot a dataset drawn from the posterior mean, to give a good legend
mock = Model.generate_replica_dataset(len(y), **pm)
axes.hist(mock, bins=bins, histtype="step", alpha=1.0, color="red", density=True,
label="Sample posterior predictions");
else:
# Take the next posterior sample a and generate a replica dataset
mock = Model.generate_replica_dataset(len(y), *Model.samples[j,:])
axes.hist(mock, bins=bins, histtype="step", alpha=alpha, color="red", density=True);
# Include the posterior mean model for comparison
yy = np.linspace(bins.min(), bins.max(), len(bins)*10)
pp = Model.sampling_distribution(yy, **pm)
axes.plot(yy, pp, linestyle="dashed", color="red", label="Posterior mean model");
axes.set_title(Model.name, fontsize=14)
axes.set_xlabel("Offset $y$ (kpc)", fontsize=12)
axes.set_ylabel("Normalized Frequency", fontsize=12)
axes.legend();
The use of transparency above means that the darkness of the red histogram lines below corresponds to the PPD at that offset.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
fig, ax = plt.subplots(1)
visual_check(ax, Model1, Nreps=100);
Based on these visual checks, would you say the model does a good job of predicting the observed data?
I_have_thought_about_the_above_question = False # change to True when true
# YOUR CODE HERE
raise NotImplementedError()
assert I_have_thought_about_the_above_question
2c. Quantitative posterior predictive model check¶
Now let's quantify the (in)adequacy of the fit with a quantitative posterior predictive model check, based on the test_statistic
function you've already defined.
To sample the posterior predictive distribution of test statistics $p[T(y_\mathrm{rep})|y,H]$, we need to generate replica datasets from the model:
def distribution_of_T(Model):
"""
Compute T(yrep) for each yrep drawn from the posterior predictive distribution,
using parameter samples stored in Model.
"""
return np.array([T(Model.generate_replica_dataset(len(y), *p)) for p in Model.samples])
We can now do the following:
- plot a histogram of $T(\mathrm{mock~data})$
- compare that distribution with $T(\mathrm{real~data})$
- compute and report the p-value for $T(\mathrm{real~data})$
And we want all of that in packaged in functions of the model, so that we can re-use it later (on different models!).
First write a function to estimate the p-value, $P(T > T(y)|y,H)$, based on posterior predictions from the chains you ran above.
def pvalue(Model):
"""
Return the posterior predictive p-value, P(T > T(y)|y,H):
"""
# YOUR CODE HERE
raise NotImplementedError()
Here's a function that plots the distribution of T, and reports the p-value:
def posterior_predictive_check(Model, nbins=25):
"""
Compute the posterior predictive distribution of the test statistic T(y_rep), and compare with T(y_obs)
"""
# Compute distribution of T(yrep):
TT = distribution_of_T(Model)
# Plot:
plt.rcParams['figure.figsize'] = (7.0, 5.0)
plt.hist(TT, bins=nbins, histtype="step", color="red", label="$P(T(y_{\\rm rep})|y)$")
# Overlay T(y_obs):
plt.axvline(x=T_true, color="gray", linestyle="dashed", label="$T(y_{\\rm observed})$")
plt.xlabel("Test statistic T(y)", fontsize=12)
plt.ylabel("Posterior predictive probability density", fontsize=12)
plt.legend();
# Compute p-value:
p = pvalue(Model)
print("p-value =", p)
return p
Let's see how it does:
p1 = posterior_predictive_check(Model1)
Your p-value will probably be quite small, of order $10^{-3}$. The precise value will vary because of all the randomization involved in estimating it (and in the data set itself), and the fact that we are out in the tail of the posterior predictive distribution.
This_checks_out_and_makes_sense = False # change to True if true
# YOUR CODE HERE
raise NotImplementedError()
assert This_checks_out_and_makes_sense
3. Evaluating a more complex model¶
Next, we'll repeat all of that (it should be much easier this time!) with a second model. To make things interesting, we'll make it a model with at least two parameters. Since this is a PDF over distances, it should be defined over the non-negative real line. The test statistic will, of course, have to be the same as before.
Note: We will discuss and agree on an alternative model, including priors, in class. You are free to experiment with other options on your own, of course, though the notebook you turn in should use the agreed model.
Include a PGM and the generative model expressions below:
Space for PGM and generative model
Copy/paste the implementation of ExponentialModel
and edit it accordingly. For grading convenience, we'll call the class AlternativeModel
, but the name
attribute set in its constructor can and should be more descriptive.
# class AlternativeModel(Model):
# ....
# YOUR CODE HERE
raise NotImplementedError()
We also need a dictionary of prior PDFs, analogous to the first model.
# alt_priors = {...}
# YOUR CODE HERE
raise NotImplementedError()
Next we instaniate it
Model2 = AlternativeModel(alt_priors)
... and test the prior sampling function:
for pars in Model2.draw_samples_from_prior(12):
print(pars)
Test the posterior function on a couple of these prior samples:
for pars in Model2.draw_samples_from_prior(5):
print('Log-posterior for', pars, '=', Model2.log_posterior(**pars))
Test the mock data generator. Here and in the following test, you can replace the call to draw_samples_from_prior
with specific parameter values if your prior is so wide that you're unlikely to see anything on the plot.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
plt.hist(Model2.generate_replica_dataset(len(y), **Model2.draw_samples_from_prior(1)[0]), bins=bins, color="lightgray");
plt.xlabel("Offset $y$ (kpc)", fontsize=12);
plt.ylabel("Frequency (replica)", fontsize=12);
Test the sampling distribution.
plt.rcParams['figure.figsize'] = (6.0, 4.0)
yy = np.linspace(bins.min(), bins.max(), len(bins)*10)
plt.plot(yy, Model2.sampling_distribution(yy, **Model2.draw_samples_from_prior(1)[0]));
plt.xlabel("Measured distance $y$", fontsize=12);
plt.ylabel("$p(y|a)$", fontsize=12);
Fit the model. Remember that a thoughtless copy/paste of the corresponding Model1 lines above will not work!
%%time
# Model2.draw_samples_from_posterior(starting_params=[...], nsteps=..., nwalkers=...)
# YOUR CODE HERE
raise NotImplementedError()
Check the convergence
# Model2.check_chains(burn=..., maxlag=...)
# YOUR CODE HERE
raise NotImplementedError()
Remove burn-in.
# Model2.remove_burnin(...)
# YOUR CODE HERE
raise NotImplementedError()
assert np.all(Model2.neff >= 1000)
Since we have multiple parameters this time around, let's look at their covariances. (The plot size below is good for 2 parameters, but increase it as needed.)
plotGTC(Model2.samples, paramNames=Model2.param_labels, figureSize=6,
customLabelFont={'size':12}, customTickFont={'size':12}, customLegendFont={'size':16});
Print the posterior mean.
print("Posterior mean parameters = ", Model2.posterior_mean())
Visually compare posterior predictions with the data, alongside those of the exponential model.
plt.rcParams['figure.figsize'] = (14.0, 5.0)
fig, ax = plt.subplots(1,2)
visual_check(ax[0], Model1, Nreps=100);
visual_check(ax[1], Model2, Nreps=100);
Finally, compute the p-value for the alternative model.
p2 = posterior_predictive_check(Model2)
Does your model appear to reproduce the data better than the exponential one? Does it have a "better" p-value?
I_have_pondered_the_above = False # change to True when true
# YOUR CODE HERE
raise NotImplementedError()
assert I_have_pondered_the_above
More to the point, is it a better model, in the sense of being justified by the data? To answer that, we need to keep in mind that a more flexible model (i.e. with more free parameters) will usually be able to fit the data more closely, even if it's only reproducing the particular realization of noise present in our specific data set. In other words, directly comparing the maximum likelihood achieved, or even the p-values, is not enough to say that the more complex model is statistically better at reproducing the features of the observed data. For that, we will need to turn from model evaluation to model comparison.
3. Compare the models using the DIC¶
Recall that the Deviance Information Criterion is given by:
$\mathrm{DIC} = D(\langle\theta\rangle) + 2p_D = \langle D(\theta) \rangle + p_D; \quad p_D = \langle D(\theta) \rangle - D(\langle\theta\rangle)$
where $\theta$ are the parameters of a model, the deviance $D(\theta)=-2\log P(\mathrm{data}|\theta)$, and averages $\langle\rangle$ are over the posterior distribution of $\theta$.
Write a function to compute this from the posterior samples we have generated for a given model.
def DIC(Model):
"""
Compute the Deviance Information Criterion for the given model.
(In a less pedagogical world, this would logically be a method of the base Model class.)
Return a tuple: (DIC,pD)
"""
# Compute the deviance D for each sample
D = -2.0*np.array([ Model.log_likelihood(*params) for params in Model.samples ])
#pD =
#DIC =
#return DIC, pD
# YOUR CODE HERE
raise NotImplementedError()
Compute the DIC for each model.
DIC1, pD1 = DIC(Model1)
print(Model1.name+':')
print("Effective number of fitted parameters =", pD1)
print("DIC =", DIC1)
DIC2, pD2 = DIC(Model2)
print(Model2.name+':')
print("Effective number of fitted parameters =", pD2)
print("DIC =", DIC2)
Do your values of $p_D$ make sense given the dimensionality of the 2 models and their priors?
Now, to interpret this, we can compare the reduction (hopefully) in the DIC of Model 2 compared with Model 1 to the Jeffreys scale (see the Model Evaluation notes). By this metric, is your second model better at explaining the data than the exponential model?
DIC1 - DIC2, np.exp(0.5*(DIC1-DIC2))
I_have_an_answer = False # change to True when true
# YOUR CODE HERE
raise NotImplementedError()
assert I_have_an_answer
4. Compare the models using the evidence¶
In general, the recommended way to compute the evidence is using a nested sampler like dynesty
. But this notebook is already getting a bit long, and we shouldn't need to be that sophisticated in this case. So, instead, we will estimate the evidence by direct Monte Carlo integration.
To do this, note that
$p(\mathrm{data}|H)=\int d\theta \, p(\mathrm{data}|\theta,H) \, p(\theta|H)$
can be approximated by averaging the likelihood over samples from the prior:
$p(\mathrm{data}|H) \approx \frac{1}{m}\sum_{k=1}^m p(\mathrm{data}|\theta_k,H)$, with $\theta_k\sim p(\theta|H)$.
This estimate is much more straightforward than trying to use samples from the posterior to calculate the evidence (which would require us to be able to normalize the posterior, which would require an estimate of the evidence, ...). But in general, and especially for large-dimensional parameter spaces, it is very inefficient because the likelihood typically is large in only a small fraction of the prior volume. Still, let's give it a try.
Write a function to draw a given number of samples from the prior and use them to calculate the evidence. To avoid numerical over/underflows, use the special scipy
function logsumexp
(which we imported directly, way at the top of the notebook) to do the sum of the likelihoods. As the name implies, this function is equivalent to log(sum(exp(...)))
, but is more numerically stable.
def log_evidence(Model, N=1000):
"""
Estimate the log-evidence for the model using N samples from the prior
"""
# YOUR CODE HERE
raise NotImplementedError()
Do a quick test to check for NaNs:
log_evidence(Model1, N=2), log_evidence(Model2, N=2)
As with any monte carlo integration, we need to be confident that we have enough samples for our estimates to be usefully accurate. In this context, each estimate needs to be accurate enough that we can tell whether they are different from one another. Get a sense for the convergence of each evidence estimate by looking at how it changes with the value of N
. If necessary, increase N
until you are sure the difference between logE1
and logE2
is meaningful.
print('Model 1:')
for Nevidence in [1, 10, 100, 1000, 10000]:
%time logE1 = log_evidence(Model1, N=Nevidence)
print("From", Nevidence, "samples, the log-evidence is", logE1)
print('Model 2:')
for Nevidence in [1, 10, 100, 1000, 10000]:
%time logE2 = log_evidence(Model2, N=Nevidence)
print("From", Nevidence, "samples, the log-evidence is", logE2)
It_looks_convincingly_converged = False # change to True when true
# YOUR CODE HERE
raise NotImplementedError()
assert It_looks_convincingly_converged
So, we have log evidences computed for each model. Now what? We can compare their difference to the Jeffreys scale again:
logE2 - logE1, np.exp(logE2 - logE1)
Note that we might end up with a different conclusion as to the strength of any preference of Model 2, compared with the DIC! The reason is that the evidence explicitly accounts for the information in the prior (which, recall, counts as part of the model definition), while the DIC does this much less directly.
We could also be good Bayesians and admit that there should be a prior distribution in model space. For example, maybe I have a very compelling theoretical reason why the offset distribution should be exponential (I don't, but just for example). Then, I might need some extra convincing that an alternative model is required.
We would then compute the ratio of the posterior probabilities of the models as follows:
prior_H1 = 0.9 # or your choice
prior_H2 = 1.0 - prior_H1 # assuming only these two options
log_post_H1 = np.log(prior_H1) + logE1
log_post_H2 = np.log(prior_H2) + logE2
print('Difference of log posteriors (H2-H1):', log_post_H2-log_post_H1)
print('Ratio of posteriors (H2/H1):', np.exp(log_post_H2-log_post_H1)) # NB this one might over/underflow
Depending on the fitness of the alternative model, you may find that only an extremely lopsided prior in model space would influence your conclusion.