Mixture models¶
[1]:
import warnings
import multiprocessing
import pandas as pd
import numpy as np
import scipy.optimize
import scipy.stats as st
import bebi103
import bokeh_catplot
import bokeh.io
bokeh.io.output_notebook()
If you are running this notebook, be sure to update the ``bebi103`` package by doing ``pip install –upgrade bebi103`` on the command line.
In a previous lesson, we explored a Negative-Binomial model for mRNA counts. The data set we used came from this paper from the Elowitz lab and featured smFISH performed in 279 cells for the genes rex1, rest, nanog, and prdm14. The data set may be downloaded here.
Multiple cell types?¶
While the Negative Binomial model could work for some genes, exploratory data analysis shows an obvious bimodality in the distribution of counts of Rex1 transcripts.
[2]:
df = pd.read_csv("../data/singer_transcript_counts.csv", comment="#")
p = bokeh_catplot.ecdf(
data=df, val="Rex1", x_axis_label="Rex1 mRNA count", conf_int=True
)
bokeh.io.show(p)
We can define a cell type by its expression profile. The bimodality of the ECDF suggests that there are more than one cell types present, since the ECDF clearly does not follow a Negative Binomial distribution.
A Negative Binomial mixture model¶
What is the story for a distribution of mRNA counts coming from more than one cell type? To be concrete, let’s say we ahve two cell types, one with one level of bursty expression of the gene of interest, and another with a different level of bursty expression. We would expect the number of mRNA transcripts to be distributed according to a linear combination of negative binomial distributions. Such a linear combination of distributions is called a mixture model. For this mixture model, we can write out the PMF for an individual measurement of \(n\) mRNA transcripts as
\begin{align} f(n\mid \alpha_1, \beta_1, \alpha_2, \beta_2, w) &= w\,\frac{(n + \alpha_1 - 1)!}{n!\,(\alpha_1-1)!}\,\left(\frac{\beta_1}{1+\beta_1}\right)^{\alpha_1}\left(\frac{1}{1+\beta_1}\right)^{n} \\[1em] &\;\;\;\;+ (1-w) \,\frac{(n + \alpha_2 - 1)!}{n!\,(\alpha_2-1)!}\,\left(\frac{\beta_2}{1+\beta_2}\right)^{\alpha_2}\left(\frac{1}{1+\beta_2}\right)^{n}. \end{align}
Where, \(w\) is the fraction of cells that are of type one (burst size and frequency are determined by \(1/\beta_1\) and \(\alpha_1\)), and \(1-w\) is the fraction of cells that are of type two. We can write this likelihood more concisely as
\begin{align} &n_i \sim w \, \text{NegBinom}(\alpha_1, \beta_1) + (1-w)\,\text{NegBinom}(\alpha_2, \beta_2) \; \forall i, \\[1em] &b_1 = 1/\beta_1, \\[1em] &b_2 = 1/\beta_2. \end{align}
(Recall from the last lesson that we are choosing to parametrize the Negative Binomial distributions with the burst size \(b\), related to \(\beta\) as \(b = 1/\beta\), so going forward we will use \(b_1\) and \(b_2\).)
MLE for a mixture model¶
Mixture models offer some unique challenges in modeling. They are revealed by looking at the log likelihood, which I write again here for a single measurement \(n\) for clarity.
\begin{align} \ell(n\mid \alpha_1, b_1, \alpha_2, b_2, w) = \ln(w\,x_1 + (1-w)x_2), \end{align}
where
\begin{align} x_i = \frac{(n + \alpha_i - 1)!}{n!\,(\alpha_i-1)!}\,\left(\frac{1}{1+b_i}\right)^{\alpha_i}\left(\frac{b_i}{1+b_i}\right)^{n}. \end{align}
The log-sum-exp trick¶
While the logarithm of a product is conveniently split, we cannot split the logarithm of a sum. If we compute the sum directly, we will get serious underflow errors for parameters for which the terms \(x_1\) or \(x_2\) are small. To compute this is a more numerically stable way, we need to use the log-sum-exp trick. To do this, we re-write the \(x_i\) as \(x_i = \mathrm{e}^{\ln x_i}\).
\begin{align} \ln f(n\mid \alpha_1, b_1, \alpha_2, b_2, w) = \ln(w\,\mathrm{e}^{\ln x_1} + (1-w)\mathrm{e}^{\ln x_2}). \end{align}
For a moment, let us assume that \(x_1 > x_2\). Then, we can factor the log likelihood as
\begin{align} \ln f(n\mid \alpha_1, b_1, \alpha_2, b_2, w) &= \ln\left[\mathrm{e}^{\ln x_1}\left(w + (1-w)\mathrm{e}^{\ln x_2 - \ln x_1}\right)\right] \\[1em] &= \ln x_1 + \ln\left(w + (1-w)\mathrm{e}^{\ln x_2 - \ln x_1}\right). \end{align}
Provided \(w\) is not too close to zero, the argument of the logarithm of the sum is now of order \(w\). So, when computing the log of the sum of weighted probabilities, we should factor out the largest term and then proceed. This is known as the log-sum-exp trick. Conveniently, this is implemented in Scipy as scipy.special.logsumexp()
. Be sure to read the docs to understand how this works,
and we will use it below in our log-likelihood calculation.
Model identifiability¶
A nonidentifiable model is a model for which we cannot unambiguously determine the parameter values. That is, two or more parameter sets are observationally equivalent. This is the case for this mixture model because of label switching. In this mixture model, it is arbitrary which \((\alpha, b)\) pair we label as \((\alpha_1, b_1)\) or \((\alpha_2, b_2)\). We can switch the labels, and also change \(w\) to \(1-w\), and we have exactly the same log likelihood. You can imaging that if we have a mixture of more than two distributions, the label switching problem gets worse, as we can have various combinations of label switching.
There is no one way to get around this kind of nonidentifiability (or many other kinds of nonidentifiability, for that matter). Since we are finding a maximum likelihood estimate, and the value of the likelihood is identical upon switching labels, we do not really need to worry about it here, though this could become a problem when computing confidence intervals, since we need to do the MLE calculation many times with different bootstrap sample data sets each time.
The log-likelihood¶
With these two considerations in mind, let’s code up the log-likelihood.
[3]:
def log_like_mix(alpha1, b1, alpha2, b2, w, n):
"""Log-likeihood of binary Negative Binomial mixture model."""
# Fix nonidentifieability be enforcing values of w
if w < 0 or w > 1:
return -np.inf
# Physical bounds on parameters
if alpha1 < 0 or alpha2 < 0 or b1 < 0 or b2 < 0:
return -np.inf
logx1 = st.nbinom.logpmf(n, alpha1, 1/(1+b1))
logx2 = st.nbinom.logpmf(n, alpha2, 1/(1+b2))
# Multipliers for log-sum-exp
lse_coeffs = np.tile([w, 1-w], [len(n), 1]).transpose()
# log-likelihood for each measurement
log_likes = scipy.special.logsumexp(np.vstack([logx1, logx2]), axis=0, b=lse_coeffs)
return np.sum(log_likes)
Initial guess¶
The optimization problem is more difficult now, since we have five parameters. We should provide the best initial guesses we can. To come up with guesses, we note in the ECDF that that the second inflection point occurs around a value of the ECDF of 0.2. We thus can guess \(w \approx 0.2\). To get guesses for \(\alpha_1\) and \(b_1\), we take the first 20% of the counts and get an MLE using the single-Negative Binomial model. We do the same for the remaining 80% of counts to get guesses for \(\alpha_2\) and \(b_2\). To do this, we need to borrow our MLE function from a previous lesson.
[4]:
def log_like_iid_nbinom(params, n):
"""Log likelihood for i.i.d. NBinom measurements, parametrized
by alpha, b=1/beta."""
alpha, b = params
if alpha <= 0 or b <= 0:
return -np.inf
return np.sum(st.nbinom.logpmf(n, alpha, 1/(1+b)))
def mle_iid_nbinom(n):
"""Perform maximum likelihood estimates for parameters for i.i.d.
NBinom measurements, parametrized by alpha, b=1/beta"""
with warnings.catch_warnings():
warnings.simplefilter("ignore")
res = scipy.optimize.minimize(
fun=lambda params, n: -log_like_iid_nbinom(params, n),
x0=np.array([3, 3]),
args=(n,),
method='Powell'
)
if res.success:
return res.x
else:
raise RuntimeError('Convergence failed with message', res.message)
We can now write a function to get an initial guess of the \(\alpha\)’s and \(b\)’s.
[5]:
def initial_guess_mix(n, w_guess):
"""Generate initial guess for mixture model."""
n_low = n[n < np.percentile(n, 100*w_guess)]
n_high = n[n >= np.percentile(n, 100*w_guess)]
alpha1, b1 = mle_iid_nbinom(n_low)
alpha2, b2 = mle_iid_nbinom(n_high)
return alpha1, b1, alpha2, b2
Let’s give it a test run.
[6]:
# Extract the values for Rex1
n = df['Rex1'].values
# Guess for w
w_guess = 0.2
# Generate initial guess
initial_guess_mix(n, w_guess)
[6]:
(2.2311645373863556, 8.874581092187775, 6.539376250892679, 25.47713820776582)
The guesses seem reasonable, so we will proceed using this method of generating initial guesses.
Performing the optimization¶
Now we’ll code up a function to generate the initial guess, and then perform the optimization. Note that below I have decreased the tolerance down to \(10^{-6}\). I found this to be necessary for this problem; the solver would be off by about 5% if I gave it similar, but different, initial guesses. The lower the tolerance, the closer to zero you require the gradient of the log likelihood to be.
[7]:
def mle_mix(n, w_guess):
"""Obtain MLE estimate for parameters for binary mixture
of Negative Binomials."""
with warnings.catch_warnings():
warnings.simplefilter("ignore")
res = scipy.optimize.minimize(
fun=lambda params, n: -log_like_mix(*params, n),
x0=[*initial_guess_mix(n, w_guess), w_guess],
args=(n,),
method='Powell',
tol=1e-6,
)
if res.success:
return res.x
else:
raise RuntimeError('Convergence failed with message', res.message)
Let’s get our MLE!
[8]:
popt = mle_mix(n, w_guess)
# Check out optimal parameters
print('α1 = {0:f}\nb1 = {1:f}\nα2 = {2:f}\nb2 = {3:f}\nw = {4:f}'.format(*popt))
α1 = 3.497013
b1 = 4.104911
α2 = 5.089623
b2 = 31.810385
w = 0.160422
Comparison to empirical CDF¶
Let’s now see how the CDFs parametrized by the MLE compare with the ECDF. The theoretical CDF for a mixture model is computed from the weights of each distribution in the mixture. If the CDF from the first Negative Binomial distribution is \(F_1(n;\alpha_1^*,b_1^*)\) and that of the second is \(F_2(n;\alpha_2^*,b_2^*)\), then the theoretical ECDF for the mixture model, parametrized by the MLE, is
\begin{align} F(n;\alpha_1^*, b_1^*, \alpha_2^*, b_2^*, w^*) = w^* F_1(n;\alpha_1^*,b_1^*) + (1-w^*) F_2(n;\alpha_2^*,b_2^*). \end{align}
[9]:
# Unpack parameters
alpha1_mle, b1_mle, alpha2_mle, b2_mle, w_mle = popt
n_theor = np.arange(0, n.max()+1)
cdf_theor = w_mle * st.nbinom.cdf(n_theor, alpha1_mle, 1/(1+b1_mle))
cdf_theor += (1 - w_mle) * st.nbinom.cdf(n_theor, alpha2_mle, 1/(1+b2_mle))
# Weave together to make staircase for discrete distribution
n_plot, cdf_plot = bebi103.viz.cdf_to_staircase(n_theor, cdf_theor)
p.line(n_plot, cdf_plot, line_color='orange', line_width=2)
bokeh.io.show(p)
We could, for comparison, also add the theoretical CDF for the Negative Binomial (non-mixture) model.
[10]:
alpha_mle, b_mle = mle_iid_nbinom(n)
cdf_theor = st.nbinom.cdf(n_theor, alpha_mle, 1/(1+b_mle))
# Weave together to make staircase for discrete distribution
n_plot, cdf_plot = bebi103.viz.cdf_to_staircase(n_theor, cdf_theor)
p.line(n_plot, cdf_plot, line_color='tomato', line_width=2)
bokeh.io.show(p)
We will discuss graphical methods of model assessment later in this lesson, but it is clear from the crude plot above that the Negative Binomial model is not commensurate with the data, but the mixture model is.
Confidence interval for the mixture model¶
Let’s now compute a confidence interval for the mixture model. To use bebi103.draw_bs_reps_mle()
, we need to first write a function to generate data out of the mixture model. To do this, we first draw a random number to see which Negative Binomial distribution to draw the count out of (either the one parametrized with \(\alpha_1\) and \(b_1\), or the one parametrized with \(\alpha_2\) and \(b_2\)). We then draw a count out of the distribution we chose.
[11]:
def gen_mix(params, size, rg):
"""Generate data for the mixture model."""
alpha1, b1, alpha2, b2, w = params
n = np.empty(size)
for i in range(size):
low_cell_type = rg.uniform() < w
if low_cell_type:
n[i] = rg.negative_binomial(alpha1, 1/(1+b1))
else:
n[i] = rg.negative_binomial(alpha2, 1/(1+b2))
return n
Now that we have this function in hand, we can generate our bootstrap samples (in parallel). This can take a while.
[12]:
bs_reps = bebi103.draw_bs_reps_mle(
mle_mix,
gen_mix,
n,
mle_args=(w_guess,),
gen_args=(),
size=5000,
n_jobs=3,
progress_bar=True,
)
100%|██████████| 1666/1666 [11:20<00:00, 2.45it/s]
100%|██████████| 1668/1668 [11:21<00:00, 2.45it/s]
100%|██████████| 1666/1666 [11:24<00:00, 2.43it/s]
Now that we have the bootstrap samples, we can compute our confidence intervals by taking percentiles. The columns are respectively for \(\alpha_1^*\), \(b_1^*\), \(\alpha_2^*\), \(b_2^*\), and \(w^*\).
[13]:
np.percentile(bs_reps, [2.5, 97.5], axis=0)
[13]:
array([[ 2.00606476, 1.605249 , 4.13641432, 25.11052257, 0.11620727],
[ 8.35914298, 8.38831565, 6.48082657, 39.09756958, 0.20872937]])
We can visualize the confidence region with a corner plot.
[14]:
# Package replicates in data frame for plotting
df_res = pd.DataFrame(data=bs_reps, columns=['α1*', 'b1*', 'α2*', 'b2*', 'w*'])
# Eliminate possible erroneous solves (delete points outside of 99.9 percentile)
inds = ((df_res < df_res.quantile(0.995)) & (df_res > df_res.quantile(0.015))).all(axis=1)
with warnings.catch_warnings():
warnings.simplefilter("ignore")
p = bebi103.viz.corner(
samples=df_res.loc[inds, :],
pars=['α1*', 'b1*', 'α2*', 'b2*', 'w*'],
show_contours=True,
levels = [0.95],
)
bokeh.io.show(p)
The corner plot of the confidence region shows that \(\alpha_1^*\) and \(b_1^*\) are strongly correlated, as are \(\alpha_2^*\) and \(b_2^*\). However, the MLEs of the parameters for the respective Negative Binomials are not correlated with each other.
A mixture model for all cells¶
If we assume a Negative Binomial (non-mixture) model for the mRNA transcript counts, we can independently perform maximum likelihood estimates for the burst frequency and burst size for each gene. However, in the mixture model, we do need to take into account the fact that the mRNA counts for the respective genes are measured in the same cells. In our modeling, that is manifest in the parameter \(w\), which specifies the relative fraction of cell types, being shared among all four genes. The generative PMF for a single set of observations \(\mathbf{n} = \{n_\mathrm{Rex1}, n_\mathrm{Rest}, n_\mathrm{Nanog}, n_\mathrm{Prdm14}\}\) is given by
\begin{align} f(\mathbf{n};\boldsymbol{\alpha}, \mathbf{b}, w) = w\prod_j x_{1j} + (1-w)\prod_j x_{2j}, \end{align}
where
\begin{align} x_{ij} = \frac{(n + \alpha_{ij} - 1)!}{n!\,(\alpha_{ij}-1)!}\,\left(\frac{1}{1+b_{ij}}\right)^{\alpha_{ij}}\left(\frac{b_{ij}}{1+b_{ij}}\right)^{n}, \end{align}
where \(i\) indexes the distribution of the mixture and \(j\) indexes the gene. We can code up the log likelihood of this by modifying our original log_like_mix()
function to take in arrays.
[15]:
def log_like_mix_all(alpha1, b1, alpha2, b2, w, n):
"""Log-likelihood of a mixuture model for
all genes. Here, `alpha1`, `b1`, `alpha2`, and `b2` are
all length 4 arrays, and `n` is 279 by 4. The
parameter `w`, however is a scalar, shared by all genes.
"""
# Fix nonidentifieability be enforcing values of w
if w < 0 or w > 1.0:
return -np.inf
# Physical bounds on parameters
if np.any(alpha1 < 0) or np.any(alpha2 < 0) or np.any(b1 < 0) or np.any(b2 < 0):
return -np.inf
# The sum is over genes
logx1 = st.nbinom.logpmf(n, alpha1, 1/(1+b1)).sum(axis=1)
logx2 = st.nbinom.logpmf(n, alpha2, 1/(1+b2)).sum(axis=1)
# Multipliers for log-sum-exp
lse_coeffs = np.tile([w, 1-w], [len(n), 1]).transpose()
# log-likelihood for each measurement
log_likes = scipy.special.logsumexp(np.vstack([logx1, logx2]), axis=0, b=lse_coeffs)
return np.sum(log_likes)
The initial guess for this model is also a bit more challenging. While we can see immediately from the Rex1 counts that the lower mRNA counts account for about 20% of the cells, this is not obvious in the others. We can, however label a “pseudo cell type” for each row of the data frame; if the Rex1 expression is below the percentile corresponding to our guess for the value of \(w\), we consider this one part of the mixture, and if it is above, it is part of the other.
[16]:
def initial_guess_mix_all(df, w_guess):
pseudo_cell_type = df["Rex1"] < np.percentile(df["Rex1"], 100*w_guess)
guess = []
for gene in df:
n = df[gene].values.astype(int)
n_low = n[pseudo_cell_type]
n_high = n[~pseudo_cell_type]
alpha1, b1 = mle_iid_nbinom(n_low)
alpha2, b2 = mle_iid_nbinom(n_high)
guess += [alpha1, b1, alpha2, b2]
return guess + [w_guess]
For speed later when we use bootstrap methods to compute the confidence intervals, we will compute this initial guess once and use it over and over to save on the four optimizations that are necessary to compute the initial guess.
[17]:
params_0 = initial_guess_mix_all(df, 0.2)
Also for speed, we should extract the counts from the data frame as a Numpy array.
[18]:
n = df.values.astype(int)
Now, we can write a function to perform the MLE. First, we need to write a function to compute the negative log likelihood, including unpacking the 17 parameters.
[19]:
def neg_log_like_all(params, n):
"""
Negative log-likelihood.
"""
n_genes = n.shape[1]
alpha1 = params[:n_genes]
b1 = params[n_genes:2*n_genes]
alpha2 = params[2*n_genes:3*n_genes]
b2 = params[3*n_genes:4*n_genes]
w = params[-1]
return -log_like_mix_all(alpha1, b1, alpha2, b2, w, n)
Finally, we can perform the optimization. Because we will use bootstrapping in a moment to compute confidence intervals for the parameters, we will encapsulate the MLE calculation in a function.
[20]:
def mle_mix_all(n, params_0):
"""Obtain MLE estimate for parameters for binary mixture
of Negative Binomials."""
with warnings.catch_warnings():
warnings.simplefilter("ignore")
res = scipy.optimize.minimize(
fun=neg_log_like_all,
x0=params_0,
args=(n,),
method='powell',
tol=1e-6,
)
if res.success:
return res.x
else:
raise RuntimeError('Convergence failed with message', res.message)
Let’s do it!
[21]:
mle = mle_mix_all(n, params_0)
# Put in a data frame for ease of reference
df_mle = pd.DataFrame(data=np.concatenate((mle, np.array([mle[-1]]*3))).reshape(5, 4),
columns=df.columns,
index=['alpha1', 'b1', 'alpha2', 'b2', 'w'])
# Take a look at the results
df_mle
[21]:
Rex1 | Rest | Nanog | Prdm14 | |
---|---|---|---|---|
alpha1 | 7.096214 | 8.891623 | 5.362176 | 0.952204 |
b1 | 24.919242 | 9.694261 | 21.176276 | 6.493398 |
alpha2 | 1.115415 | 3.489837 | 1.174485 | 0.440116 |
b2 | 44.837432 | 14.026622 | 18.588349 | 1.664736 |
w | 0.697829 | 0.697829 | 0.697829 | 0.697829 |
The MLEs converged on \(w \approx 0.7\), which corresponds to \(w \approx 0.3\) from before (we encountered label switching in this optimization). We can do a quick graphical assessment of the model.
[22]:
def plot_theor_cdf(gene, p):
"""Add a theoretical CDF to plot"""
n_theor = np.arange(0, df[gene].max()+1)
s = df_mle[gene]
cdf_theor = s['w'] * st.nbinom.cdf(n_theor, s['alpha1'], 1/(1+s['b1']))
cdf_theor += (1 - s['w']) * st.nbinom.cdf(n_theor, s['alpha2'], 1/(1+s['b2']))
# Weave together to make staircase for discrete distribution
n_plot, cdf_plot = bebi103.viz.cdf_to_staircase(n_theor, cdf_theor)
p.line(n_plot, cdf_plot, color='orange', line_width=2)
return p
# Make a list of plots
plots = [
bokeh_catplot.ecdf(
data=df,
val=gene,
frame_height=125,
frame_width=250,
conf_int=True,
title=gene,
)
for gene in df
]
# Add CDFs
for i, gene in enumerate(df):
plots[i] = plot_theor_cdf(gene, plots[i])
# Show them as a grid
bokeh.io.show(bokeh.layouts.gridplot(plots, ncols=2))
The mixture model seems to fit all four reasonably well, with the theoretical CDF parametrized by the MLE mostly falling within the 95% confidence interval of the ECDF. (We will explore other ways to visualize and assess models in the next parts of this lesson.) However, we note that the Rex1 data do see some deviations on the tail of the distribution for small counts.
Confidence intervals for the mixture model for all genes¶
We can attempt to compute confidence regions for the MLEs for this model for all four genes. We again start by writing a function to generate the data from the mixture model.
[23]:
def gen_mix_all(params, size, rg):
"""Generate data for the mixture model."""
alpha1 = params[:4]
b1 = params[4:8]
alpha2 = params[8:12]
b2 = params[12:16]
w = params[-1]
n = np.empty((size, 4))
for i in range(size):
low_cell_type = rg.uniform() < w
if low_cell_type:
n[i] = rg.negative_binomial(alpha1, 1/(1+b1))
else:
n[i] = rg.negative_binomial(alpha2, 1/(1+b2))
return n
We can now draw our bootstrap replicates. I will only draw six replicates for now, though, and the reason will be clear when we see the result.
[24]:
bs_mle_reps = bebi103.draw_bs_reps_mle(
mle_mix_all,
gen_mix_all,
n,
mle_args=(params_0,),
gen_args=(),
size=6,
n_jobs=3,
)
bs_mle_reps
---------------------------------------------------------------------------
RemoteTraceback Traceback (most recent call last)
RemoteTraceback:
"""
Traceback (most recent call last):
File "/Users/bois/opt/anaconda3/lib/python3.7/multiprocessing/pool.py", line 121, in worker
result = (True, func(*args, **kwds))
File "/Users/bois/opt/anaconda3/lib/python3.7/multiprocessing/pool.py", line 47, in starmapstar
return list(itertools.starmap(args[0], args[1]))
File "/Users/bois/Dropbox/git/bebi103/bebi103/utils.py", line 615, in _draw_bs_reps_mle
for _ in iterator
File "/Users/bois/Dropbox/git/bebi103/bebi103/utils.py", line 615, in <listcomp>
for _ in iterator
File "<ipython-input-20-3582a3e2aeb2>", line 18, in mle_mix_all
raise RuntimeError('Convergence failed with message', res.message)
RuntimeError: ('Convergence failed with message', 'Maximum number of function evaluations has been exceeded.')
"""
The above exception was the direct cause of the following exception:
RuntimeError Traceback (most recent call last)
<ipython-input-24-5dd5531ff3c5> in <module>
6 gen_args=(),
7 size=6,
----> 8 n_jobs=3,
9 )
10
~/Dropbox/git/bebi103/bebi103/utils.py in draw_bs_reps_mle(mle_fun, gen_fun, data, mle_args, gen_args, size, n_jobs, progress_bar, rg)
698
699 with multiprocessing.Pool(n_jobs) as pool:
--> 700 result = pool.starmap(_draw_bs_reps_mle, arg_iterable)
701
702 return np.concatenate(result)
~/opt/anaconda3/lib/python3.7/multiprocessing/pool.py in starmap(self, func, iterable, chunksize)
274 `func` and (a, b) becomes func(a, b).
275 '''
--> 276 return self._map_async(func, iterable, starmapstar, chunksize).get()
277
278 def starmap_async(self, func, iterable, chunksize=None, callback=None,
~/opt/anaconda3/lib/python3.7/multiprocessing/pool.py in get(self, timeout)
655 return self._value
656 else:
--> 657 raise self._value
658
659 def _set(self, i, obj):
RuntimeError: ('Convergence failed with message', 'Maximum number of function evaluations has been exceeded.')
We got a runtime error that the solver did not converge. This can happen; finding MLEs for a large number of parameters is very difficult and can often lead to this kind of failure. We will therefore not pursue confidence intervals for this more complicated mixture model.
That said, given the difficulty in finding MLEs for complicated models, we can’t be too sure that the single maximum likelihood estimate we got for the nineteen parameters of this model is even a good one. This should raise a reg flag for you. MLEs are very difficult to obtain for high-dimensional models, and when they are obtained, it is very difficult to verify that they do, in fact, give the maximum likelihood estimate.
We will deal with complex models and ways of investigating the failure modes in finding parameter estimates next term. For now, we will proceed by considering only the single-gene case of Rex1.
The expectation-maximization algorithm¶
I close this part of the lesson by briefly mentioning that the expectation-maximization (EM) algorithm is a commonly used effective algorithm for performing maximum likelihood estimates for mixture models. It alleviates some (but by no means all) of the problems for finding MLEs of complex models. We will not discuss it, but you should be aware of it.
Computing environment¶
[25]:
%load_ext watermark
%watermark -v -p numpy,pandas,scipy,bokeh,bokeh_catplot,bebi103,jupyterlab
CPython 3.7.5
IPython 7.9.0
numpy 1.17.4
pandas 0.24.2
scipy 1.3.1
bokeh 1.4.0
bokeh_catplot 0.1.6
bebi103 0.0.45
jupyterlab 1.2.3