The 4 Parameter Logistic (4PL) model has been recognized as a major tool to analyze the relationship between doses and responses in pharmacological experiments. A main strength of this model is that each parameter contributes an intuitive meaning enhancing interpretability of a fitted model. However, implementing the 4PL model using conventional statistical software often encounters numerical errors. This paper highlights the issue of convergence failure and presents several causes with solutions. These causes include outliers and a non-logistic data shape, so useful remedies such as robust estimation, outlier diagnostics and constrained optimization are proposed. These features are implemented in a new R package dr4pl (Dose-Response analysis using the 4 Parameter Logistic model) whose code examples are presented as a separate section. Our R package dr4pl is shown to work well for data sets where the traditional dose-response modelling packages drc and nplr fail.
Dose-response models are a primary tool of pharmacological and toxicological studies. After fitting a suitably chosen model to experimental data, the fitted parameters are used to evaluate the potency of a drug or compare the potencies of different drugs. Based on outcomes of the analysis, drug candidates may be advanced to further development. One of the most popular models among dose-response models is the 4 Parameter Logistic (4PL) model whose detailed explanation is given in Subsection 2.1. In this paper, we focus on some important challenges in the 4PL model and present methods to handle those challenges.
Fitting dose-response models to data is usually a nonlinear regression problem. One of the problems with nonlinear regression is that numerical errors often occur during a model fitting optimization process. The 4PL model is not an exception. Examples of problem data sets and diagnostics of their problems are presented in Subsection 2.2.
In this paper, we analyze the convergence failure problem within the context of the 4PL model. Descriptions of the 4PL model and existing methods are presented in Section 2. Problems are clearly illustrated using the contour of the likelihood function, shown in Section 3. This suggests a useful set of constraints to put on the parameter space, and their effectiveness is analyzed in that section as well. Non-convergence is indicated by our R package dr4pl (Landis et al. 2018) when crossing one of these boundaries, in which case robust parameter estimates are returned as described in Section 4. Section 5 presents a new method of selecting initial parameter estimates which is not currently implemented in other dose-response modelling R packages. This method will be shown to be effective in handling a non-logistic data shape in that section. Walk-through of R functions in dr4pl with real-world data sets will be presented in Section 6.
This section describes the 4PL model and existing methods that have been implemented for the 4PL model. Example data sets that motivate our research are presented as well.
The response in a pharmacological process usually slowly trends downward
at low dose levels, then rapidly drops at a certain dose level and
finally stabilizes. Tracking this sort of change in response results in
a reverse S shaped curve motivating sigmoidal curves as primary models
of interest. Robertson (1908) suggested that the biochemical nature of
such processes is well modelled by the differential equation
A major strength of the 4PL model is that each of the four parameters
has an intuitive meaning. The parameters
The effects of the IC50 parameter
As mentioned in Section 1, dose-response models are
prone to numerical errors in their optimization processes like other
nonlinear regression models; see Bates and Watts (1988) and Seber and Wild (1989) for
theoretical and practical issues related to nonlinear regression models.
The specific type of errors in dose-response models depends on the data
and statistical software at hand, and here we focus on R packages
developed for bioassay data. The R package
drc (Ritz et al. 2005) provided
various non-linear regression models for bioassay data analysis. The R
package nplr (Commo and Bot 2016)
focused on the family of errorm
to control whether an error or a warning
message is drawn when convergence failure happens, and nplr provides a
weighting scheme that reduces the effect of outliers. Both methods did
not suffice to prevent convergence failure as shown in the following.
Figure 2 shows four real-world data sets. The R
package drc fails to converge on all these four data sets and the R
package nplr fails on the last data set. In each plot of the figure,
the
The data in the lower two plots have a different problem in addition to
outliers. They seem to have insufficient data points on the right side
of their IC50 parameters. This situation is well known in dose-response
experiments in cases where solubility issues limit the maximal drug
concentration that can be used in dose titration. Error Case 3 in the
lower left plot shows two strong outliers with abnormally small response
values at the dose level 10. In addition, the response starts decreasing
at the dose levels 10 and 100 but no more responses are measured at
higher dose levels. This is because the drug fell out of solution at
higher concentrations in this experiment. We call this type of problem a
support problem because of a lack of data points on the right side,
and possibly also on the left side, of the IC50 parameter. This lack in
Error Case 3 can result in instability of an optimization process due to
difficulty in estimation of the IC50 parameter
Convergence failure of numerical optimization methods can be overcome in various ways. Experimenters often try other starting parameter values or adopt a simpler mathematical model. However, both solutions have a drawback in dose-response modelling since they require arbitrary decisions by experimenters. This limits their utility for automated, high-throughput screening and may be cost-prohibitive depending on the expense per data point. It can be foolhardy to decide which starting values should be tried next unless problems with data are diagnosed. Performing a grid search on the parameter space can be an alternative as claimed in Wang et al. (2010), but even this does not fundamentally solve the problems with the data sets in Figure 2. Shifting to a simpler model can result in loss of interpretability of the 4PL model. As pointed out in Di Veroli Giovanni et al. (2015) and Lim (2015), the 4PL model conveys useful intuitive content through their parameters.
The 4PL model given in Equation ((2)) is fitted to a
data set
To find a reason behind the convergence failure, we draw contour plots
of the sum-of-squares loss function given in Equation
((3)). To visualize the loss function
The primary resolution of the convergence failure problem comes from
imposing appropriate boundaries on the parameter space to prevent
parameter estimates from diverging to infinity or converging to 0. These
bounds should be broad enough to contain the global minimum of the loss
function when it has any local minimum. To obtain reasonable sizes of
bounds on the parameters, we focus on the Hill plot. Recall from
Equation ((2)) that the mean response is modelled as
To obtain reasonable bounds on the parameters
However, using these bounds in optimization have limitations in practice
in that they may miss important nonlinear structure inherent in the 4PL
model. In real data analysis, such confidence intervals imposed too
narrow bounds on the parameters and hindered optimization processes from
convergence even when data have good logistic shapes. Section 5.1 of
Seber and Wild (1989) present approximate confidence intervals of the true
parameters of nonlinear regression models based on the derivatives of
the mean response function
Based on this asymptotic theory, the approximate confidence intervals
for the true parameters are given in Section 5.1 of Seber and Wild (1989) as
One can wonder if bounds on the upper asymptote
Theorem 1. Let
Proof 1. See Appendix.
In this subsection, we evaluate whether the Hill bounds are appropriate
as a criterion for convergence failure. We simulate data following the
4PL model and check whether the Hill bounds prevent optimization
processes from finding the global minimum of the loss function
Figure 4 shows parameter settings of our
simulation. The left plot shows nine dose-response curves that generate
data whose values of parameters are shown in Table
1. The levels of slope parameter
Model | Parameters |
Model | Parameters |
---|---|---|---|
1 | (100,0.01,-0.5,0) | 6 | (100,0.1,-1.5,0) |
2 | (100,0.01,-1,0) | 7 | (100,1,-0.5,0) |
3 | (100,0.01,-1.5,0) | 8 | (100,1,-1,0) |
4 | (100,0.1,-0.5,0) | 9 | (100,1,-1.5,0) |
5 | (100,0.1,-1,0) |
Parameters | |||
---|---|---|---|
0.054 | 0.014 | 0 | |
0 | 0 | 0.002 | |
0.181 | 0 | 0 |
Table 2 shows the proportions of times that
the Hill bounds miss true parameters of the nine models given in Figure
4. For each simulated data set, we counted
the number of times that the Hill bounds miss any of the IC50 and slope
parameters. Note from the table that the Hill bounds contain the true
parameters for most of the simulated data. The only exception is when
data have a severe support problem
Table 3 presents the proportion of
times the Hill bounds are hit at the end of optimization processes out
of 1,000 simulations. The final parameter estimates are inside the Hill
bounds for most cases. This indicates that the Hill bounds are not too
restrictive for optimization procedures to find a local minimum of the
loss function
Parameters | |||
---|---|---|---|
0.038 | 0.009 | 0.1 | |
0.12 | 0.002 | 0.195 | |
0.398 | 0.258 | 0.079 |
As seen in the previous section, it seems that the convergence failure
results from either outliers or the support problem. We tackle each of
the causes with a different approach. Outlier problems are often dealt
with by robust estimation methods. The paper Riazoshams et al. (2009) reviewed
various methods of outlier detection in the context of nonlinear
regression, especially when the form of the mean response function ROUT
method therein
as a default outlier detection method. The indices of outliers in data
and a scatter plot with a robust fit are provided to users to enhance
their understanding of problems.
As noted in Subsection 2.2,
one of the main reasons for convergence failure is strong outliers. In
this case, the contour of the loss function
To determine which data observations are outliers, we use the method
given in Motulsky and Brown (2006). Define the
Fit a dose-response model to data using a robust regression model to
obtain residuals
Compute the median absolute deviation (MAD) of absolute valued
residuals
Rank the absolute values of the residuals from the smallest to the
largest, so that
Loop from
Compute
Compute the two-tailed p-value
If any of
Contrary to the suggestion of Motulsky and Brown (2006) to use their quantile based
measure of scale,
For robust regression, we adopt M-estimators. The letter M implies
that the M-estimators are obtained from equations that are in similar
forms with maximum likelihood estimators. The M-estimators are
obtained by minimizing
Figure 5 shows the shape of the
Using M-estimators in nonlinear regression has a few problems. First of all, the conditions for M-estimators to have asymptotic normality are complicated and difficult to be verified as mentioned in Section 12.3 of Seber and Wild (1989). This prevents us from obtaining approximate confidence intervals of true parameters based on M-estimators and their standard errors. On top of that, there are no known general results about consistency of M-estimators in nonlinear regression, which hinders us from trusting fitted parameter estimates by M-estimators. However, the absolute loss function was shown in Oberhofer (1982) to exhibit weak consistency in a nonlinear regression setting. Hence, we set the absolute loss estimator to be a default robust estimation method in our R package when the Hill bounds are attained.
Figure 6 shows the results of applying
the R package dr4pl to the error cases of Figure
2. Unlike the R packages drc and nplr, the
package dr4pl succeeded in fitting a robust regression model to all
four data sets. The
The next cause of the convergence failure is the support problem whose
possible solution is presented in this section. Final parameter
estimates of the 4PL model obtained by an optimization procedure depend
on initial estimates supplied to the procedure. The traditional R
packages drc and nplr use the logistic method which computes the
initial parameter estimates as
The paper Wang et al. (2010) presented a grid algorithm for fitting the 4PL model
to many data sets generated by a high throughput screening experiment. A
main idea of their research is that fitting the Hill equation in
Equation ((6)) with multiple initial values of the
upper and lower asymptotes,
The basic idea of choosing grids resembles the idea of the Hill bounds
in that it uses asymptotic normality of parameter estimators presented
in Equation ((10)). Since we use
An advanced method of obtaining initial parameter estimates is illustrated in Subsection 8.3.2 of Ratkowsky (1983). The method is basically a variant of the method given in Mead (1970) which used a weighted least squares criterion to obtain an approximate solution of the ordinary least squares problem. We call this Mead’s method in this paper. Unlike the logistic method, Mead’s method estimates the upper asymptote and the IC50 parameter from a linear model by first fixing the slope and lower asymptote. Since Mead’s method performs a grid search on the slope parameter, it can better obtain initial parameter estimates when there are support problems like those in the lower row of Figure 2. In such cases, obtaining a proper initial estimate of the slope parameter is difficult and performing a grid search for it can be useful.
Mead’s method starts from a model
To find estimators of
As pointed out in the previous subsection, Mead’s method is expected to find better initial parameter estimates than the logistic method when data have the support problem. By simulating data and fitting 4PL models with different initialization techniques, we can see whether the Mead method actually results in final fitted models with lower loss values than the logistic method. We also compare the two methods when data do not exhibit a support problem. By doing that, we can get insight into their relative strengths under various situations.
In this simulation study, we generate data from the models of Subsection 3.3 whose mean response curves are presented in the left plot of Figure 4 and whose parameter values are given in Table 2. As mentioned in that subsection, Models 1, 2 and 3 represent models without any support problem, while Models 7, 8 and 9 represent severe support problems. Models 4, 5 and 6 exhibit mild levels of support problems. For each model, 1,000 data sets are generated and the standard deviation of normally distributed errors was fixed at 5.
Table 4 shows the numbers of times that
the Mead method outperforms the logistic method in the sense that it has
a lower loss value for final fitted models out of 1,000 simulations. The
top row of the table represents Models 1, 2 and 3 from left to right. As
can be seen from the top row, the Mead method generally performs worse
than the logistic method except when the slope is very steep
Parameters | |||
---|---|---|---|
442 | 443 | 697 | |
651 | 568 | 682 | |
661 | 767 | 843 |
In this section, we walk through example R code in which dose-response data are analyzed using our R package dr4pl. By following steps shown below, users can understand how to use dr4pl to handle data sets that have the convergence failure issue and what kind of detailed analysis can be done by setting options of R functions in dr4pl.
We start with Error Case 1 shown in Figures 2
and 6. As can be seen in those figures,
this data set contains an extreme outlier at the dose level
> library(drc)
>
> drm(Response~Dose, data = drc_error_1, fct = LL.4())
Error in optim(startVec, opfct, hessian = TRUE, method = optMethod, control =
list(maxit = maxIt, : non-finite finite-difference value [4]
Error in drmOpt(opfct, opdfct1, startVecSc, optMethod, constrained, warnVal, :
Convergence failed
Note that this error message indicates convergence failure but no
explanation about its cause is given. On the contrary, fitting the 4PL
model with dr4pl returns an object of the class "dr4pl"
which
contains useful information about the convergence failure and a possible
solution.
> library(dr4pl)
>
> dr4pl.error.1 <- dr4pl(Response~Dose, data = drc_error_1, method.init = "logistic",
+ use.Hessian = TRUE)
> class(dr4pl.error.1)
[1] "dr4pl"
> dr4pl.error.1$convergence # TRUE indicates convergence success,
+ # FALSE indicates convergence failure.
[1] FALSE
In this code, some input arguments such as method.init
and
use.Hessian
are specified as different values from the default values
of dr4pl. For example, the default value of method.init
is Mead
but it is specified as logistic
in the code. This is to ensure that
the function dr4pl
works with the same parameter setting as drc. The
R package drc uses the logistic method as its default initialization
method and the Hessian matrix in its confidence interval estimation.
Note that the function dr4pl
does not throw an error but returns the
member dr4pl.error.1$convergence
which indicates that convergence
failure happens during the optimization process.
As a possible solution to the convergence failure, the object
dr4pl.error.1
provides a member object dr4pl.robust
that results
from robust estimation and the outlier detection algorithm explained in
Subsection 4.1. We call
the object dr4pl.error.1$dr4pl.robust
the robust child object of its
parent object dr4pl.error.1
.
> dr4pl.robust.1 <- dr4pl.error.1$dr4pl.robust
> class(dr4pl.error.1$dr4pl.robust)
[1] "dr4pl"
> dr4pl.robust.1$convergence
[1] TRUE
It can be seen from these code lines that the robust child object does
not report the convergence failure issue. To check results of robust
estimation and outlier detection, we check the plot of the object
dr4pl.robust.1
.
> dr4pl.robust.1$idx.outlier
[1] 132 140 127 134 133 137 141 139 130 129 128 102
> plot(dr4pl.robust.1, indices.outlier = dr4pl.robust.1$idx.outlier)
The member idx.outlier
of the object dr4pl.robust.1
shows the
indices of the outliers in the input data frame drc_error_1
detected
by our algorithm. It can be seen that the last sentence in this code
generates the same plot as the upper left plot of Figure
6. As mentioned in Subsection
4.1, the absolute loss is
given as the default robust estimation method. This implies that
adopting the absolute loss successfully resolves the convergence failure
problem of Error Case 1.
Our R package dr4pl provides other real-world data sets than Error
Cases 1 - 4. We analyze the data set sample_data_5
to compare the two
different initialization methods, the logistic and Mead methods, which
were presented in Section 5.
> dr4pl.logistic.5 <- dr4pl(Response~Dose,
+ data = sample_data_5,
+ method.init = "logistic")
> ggplot.logistic <- plot(dr4pl.logistic.5, text.title = "Logistic method")
>
> dr4pl.Mead.5 <- dr4pl(Response~Dose,
+ data = sample_data_5)
> ggplot.Mead <- plot(dr4pl.Mead.5, text.title = "Mead's method")
>
> grid.arrange(ggplot.logistic, ggplot.Mead, nrow = 1, ncol = 2)
Running this code generates the two plots in Figure 7. Note that outliers are not denoted by red triangles in these plots since they are generated by the parent objects but not by their robust child objects. It can be seen from the figure that the data set has the support problem in the sense that it is hard to determine an IC50 parameter value visually and data points on the right side of possible IC50 parameter values are not sufficiently provided. Mead’s method shown in the right plot seems to yield a closer fit to the right-most data point. This coincides with our explanation in Subsection 5.2 that the method can result in better final parameter estimates particularly when data suffer from the support problem.
The R function summary
of the R package dr4pl shows the 95%
asymptotic confidence intervals of the final parameter estimates of a
"dr4pl"
object. The following code lines show example output of
summary
applied to the data set sample_data_3
.
> dr4pl.3 <- dr4pl(Response~Dose, data = sample_data_3)
> summary(dr4pl.3)
$call
dr4pl.formula(formula = Response ~ Dose, data = sample_data_3)
$coefficients
Estimate StdErr 2.5 % 97.5 %
UpperLimit 59858.0700194 1.328219e-01 58678.1496415 61037.9903974
IC50 5.0794918 1.531608e-05 3.7133017 6.9483278
Slope -0.6117371 1.440712e-05 -0.7397224 -0.4837518
LowerLimit 980.8205686 1.354448e-01 -222.4002527 2184.0413899
attr(,"class")
[1] "summary.dr4pl"
Recall from Subsection 2.1 that the IC50 is the
dose level whose response value is halfway between the upper and lower
asymptotes
Like the R function ED
of drc and the R function getEstimates
of
nplr, the R package dr4pl implements the function IC
to compute
inhibitory concentration parameters IC
to compute
sample_data_3
.
> IC(dr4pl.3, c(10, 30, 50, 70, 90))
InhibPercent:10 InhibPercent:30 InhibPercent:50 InhibPercent:70 InhibPercent:90
184.3784198 20.2930770 5.0794918 1.2714305 0.1399363
The 4PL model has often suffered from convergence failure problems preventing experimenters from obtaining fitted 4PL models to data sets. Without diagnosing problems with data sets, experimenters should often try many different initial parameter estimates, which still does not guarantee stable convergence of optimization processes. In this paper, we presented a novel method based on studying convergence failure problems. By imposing proper conditions on the parameter space of the 4PL models, our new R package dr4pl reports convergence failure during an optimization process and presents possible resolutions to users.
Another important issue can be memory usage and speed of dr4pl, relative to other R packages nplr and drc. Measuring time and memory consumed by the three R packages for the 13 data sets provided in dr4pl showed that nplr is much slower, and used much more memory, than either of the other two methods. The enhanced features of dr4pl resulted in its being only slightly slower, with somewhat more memory usage than drc.
Asymptotic confidence intervals of M-estimators have been developed in Lim et al. (2013). An interesting open problem is the numerical implementation of these in the 4PL context. Another area of potential future extension is modelling interaction between different drugs. The R packages drc and MixLow support functionality for drug interaction modelling. While we are not aware of data sets reported in the literature that cause the convergence failure problem when there are more than one drug, it seems that outliers or the support problem may cause instability of optimization processes in those cases as well. Analysis of the Hill bounds can be a remedy for those, which is an important area of future research.
Let
Agriculture, ChemPhys, Pharmacokinetics
This article is converted from a Legacy LaTeX article using the texor package. The pdf version is the official version. To report a problem with the html, refer to CONTRIBUTE on the R Journal homepage.
Text and figures are licensed under Creative Commons Attribution CC BY 4.0. The figures that have been reused from other sources don't fall under this license and can be recognized by a note in their caption: "Figure from ...".
For attribution, please cite this work as
An, et al., "dr4pl: A Stable Convergence Algorithm for the 4 Parameter Logistic Model", The R Journal, 2019
BibTeX citation
@article{RJ-2019-003, author = {An, Hyowon and Landis, Justin T. and Bailey, Aubrey G. and Marron, J. S. and Dittmer, Dirk P.}, title = {dr4pl: A Stable Convergence Algorithm for the 4 Parameter Logistic Model}, journal = {The R Journal}, year = {2019}, note = {https://rjournal.github.io/}, volume = {11}, issue = {2}, issn = {2073-4859}, pages = {171-190} }