% File src/library/stats/man/nls.Rd % Part of the R package, https://www.R-project.org % Copyright 1995-2021 R Core Team % Distributed under GPL 2 or later \name{nls} \alias{nls} \title{Nonlinear Least Squares} \description{ Determine the nonlinear (weighted) least-squares estimates of the parameters of a nonlinear model. } \usage{ nls(formula, data, start, control, algorithm, trace, subset, weights, na.action, model, lower, upper, \dots) } \arguments{ \item{formula}{a nonlinear model \link{formula} including variables and parameters. Will be coerced to a formula if necessary.} \item{data}{an optional data frame in which to evaluate the variables in \code{formula} and \code{weights}. Can also be a list or an environment, but not a matrix.} \item{start}{a named list or named numeric vector of starting estimates. When \code{start} is missing (and \code{formula} is not a self-starting model, see \code{\link{selfStart}}), a very cheap guess for \code{start} is tried (if \code{algorithm != "plinear"}). } \item{control}{an optional \code{\link{list}} of control settings. See \code{\link{nls.control}} for the names of the settable control values and their effect.} \item{algorithm}{character string specifying the algorithm to use. The default algorithm is a Gauss-Newton algorithm. Other possible values are \code{"plinear"} for the \I{Golub}-\I{Pereyra} algorithm for partially linear least-squares models and \code{"port"} for the \sQuote{\I{nl2sol}} algorithm from the Port library -- see the references. Can be abbreviated.} \item{trace}{logical value indicating if a trace of the iteration progress should be printed. Default is \code{FALSE}. If \code{TRUE} the residual (weighted) sum-of-squares, the convergence criterion and the parameter values are printed at the conclusion of each iteration. Note that \code{\link{format}()} is used, so these mostly depend on \code{\link{getOption}("digits")}. When the \code{"plinear"} algorithm is used, the conditional estimates of the linear parameters are printed after the nonlinear parameters. When the \code{"port"} algorithm is used the objective function value printed is half the residual (weighted) sum-of-squares.} \item{subset}{an optional vector specifying a subset of observations to be used in the fitting process.} \item{weights}{an optional numeric vector of (fixed) weights. When present, the objective function is weighted least squares.} \item{na.action}{a function which indicates what should happen when the data contain \code{NA}s. The default is set by the \code{na.action} setting of \code{\link{options}}, and is \code{\link{na.fail}} if that is unset. The \sQuote{factory-fresh} default is \code{\link{na.omit}}. Value \code{\link{na.exclude}} can be useful.} \item{model}{logical. If true, the model frame is returned as part of the object. Default is \code{FALSE}.} \item{lower, upper}{vectors of lower and upper bounds, replicated to be as long as \code{start}. If unspecified, all parameters are assumed to be unconstrained. Bounds can only be used with the \code{"port"} algorithm. They are ignored, with a warning, if given for other algorithms.} \item{\dots}{Additional optional arguments. None are used at present.} } \details{ An \code{nls} object is a type of fitted model object. It has methods for the generic functions \code{\link{anova}}, \code{\link{coef}}, \code{\link{confint}}, \code{\link{deviance}}, \code{\link{df.residual}}, \code{\link{fitted}}, \code{\link{formula}}, \code{\link{logLik}}, \code{\link{predict}}, \code{\link{print}}, \code{\link{profile}}, \code{\link{residuals}}, \code{\link{summary}}, \code{\link{vcov}} and \code{\link{weights}}. Variables in \code{formula} (and \code{weights} if not missing) are looked for first in \code{data}, then the environment of \code{formula} and finally along the search path. Functions in \code{formula} are searched for first in the environment of \code{formula} and then along the search path. Arguments \code{subset} and \code{na.action} are supported only when all the variables in the formula taken from \code{data} are of the same length: other cases give a warning. Note that the \code{\link{anova}} method does not check that the models are nested: this cannot easily be done automatically, so use with care. } \section{Warning}{ \bold{The default settings of \code{nls} generally fail on artificial \dQuote{zero-residual} data problems.} The \code{nls} function uses a relative-offset convergence criterion that compares the numerical imprecision at the current parameter estimates to the residual sum-of-squares. This performs well on data of the form \deqn{y=f(x, \theta) + \varepsilon}{y = f(x, \theta) + eps} (with \eqn{var(\varepsilon) > 0}{var(eps) > 0}). It fails to indicate convergence on data of the form \deqn{y = f(x, \theta)} because the criterion amounts to comparing two components of the round-off error. To avoid a zero-divide in computing the convergence testing value, a positive constant \code{scaleOffset} should be added to the denominator sum-of-squares; it is set in \code{control}, as in the example below; this does not yet apply to \code{algorithm = "port"}. The \code{algorithm = "port"} code appears unfinished, and does not even check that the starting value is within the bounds. Use with caution, especially where bounds are supplied. } \value{ A list of \item{m}{an \code{nlsModel} object incorporating the model.} \item{data}{the expression that was passed to \code{nls} as the data argument. The actual data values are present in the \code{\link{environment}} of the \code{m} components, e.g., \code{environment(m$conv)}.} \item{call}{the matched call with several components, notably \code{algorithm}.} \item{na.action}{the \code{"na.action"} attribute (if any) of the model frame.} \item{dataClasses}{the \code{"dataClasses"} attribute (if any) of the \code{"terms"} attribute of the model frame.} \item{model}{if \code{model = TRUE}, the model frame.} \item{weights}{if \code{weights} is supplied, the weights.} \item{convInfo}{a list with convergence information.} \item{control}{the control \code{list} used, see the \code{control} argument.} \item{convergence, message}{for an \code{algorithm = "port"} fit only, a convergence code (\code{0} for convergence) and message. To use these is \emph{deprecated}, as they are available from \code{convInfo} now. } } \note{Setting \code{warnOnly = TRUE} in the \code{control} argument (see \code{\link{nls.control}}) returns a non-converged object (since \R version 2.5.0) which might be useful for further convergence analysis, \emph{but \bold{not} for inference}. } \references{ Bates, D. M. and Watts, D. G. (1988) \emph{Nonlinear Regression Analysis and Its Applications}, Wiley Bates, D. M. and Chambers, J. M. (1992) \emph{Nonlinear models.} Chapter 10 of \emph{Statistical Models in S} eds J. M. Chambers and T. J. Hastie, Wadsworth & Brooks/Cole. \url{https://netlib.org/port/} for the Port library documentation. } \author{ Douglas M. Bates and Saikat DebRoy: David M. Gay for the Fortran code used by \code{algorithm = "port"}. } \seealso{ \code{\link{summary.nls}}, \code{\link{predict.nls}}, \code{\link{profile.nls}}. Self starting models (with \sQuote{automatic initial values}): \code{\link{selfStart}}. } \examples{ \dontshow{od <- options(digits=5)} require(graphics) DNase1 <- subset(DNase, Run == 1) ## using a selfStart model fm1DNase1 <- nls(density ~ SSlogis(log(conc), Asym, xmid, scal), DNase1) summary(fm1DNase1) ## the coefficients only: coef(fm1DNase1) ## including their SE, etc: coef(summary(fm1DNase1)) ## using conditional linearity fm2DNase1 <- nls(density ~ 1/(1 + exp((xmid - log(conc))/scal)), data = DNase1, start = list(xmid = 0, scal = 1), algorithm = "plinear") summary(fm2DNase1) ## without conditional linearity fm3DNase1 <- nls(density ~ Asym/(1 + exp((xmid - log(conc))/scal)), data = DNase1, start = list(Asym = 3, xmid = 0, scal = 1)) summary(fm3DNase1) ## using Port's nl2sol algorithm fm4DNase1 <- nls(density ~ Asym/(1 + exp((xmid - log(conc))/scal)), data = DNase1, start = list(Asym = 3, xmid = 0, scal = 1), algorithm = "port") summary(fm4DNase1) ## weighted nonlinear regression Treated <- Puromycin[Puromycin$state == "treated", ] weighted.MM <- function(resp, conc, Vm, K) { ## Purpose: exactly as white book p. 451 -- RHS for nls() ## Weighted version of Michaelis-Menten model ## ---------------------------------------------------------- ## Arguments: 'y', 'x' and the two parameters (see book) ## ---------------------------------------------------------- ## Author: Martin Maechler, Date: 23 Mar 2001 pred <- (Vm * conc)/(K + conc) (resp - pred) / sqrt(pred) } Pur.wt <- nls( ~ weighted.MM(rate, conc, Vm, K), data = Treated, start = list(Vm = 200, K = 0.1)) summary(Pur.wt) ## Passing arguments using a list that can not be coerced to a data.frame lisTreat <- with(Treated, list(conc1 = conc[1], conc.1 = conc[-1], rate = rate)) weighted.MM1 <- function(resp, conc1, conc.1, Vm, K) { conc <- c(conc1, conc.1) pred <- (Vm * conc)/(K + conc) (resp - pred) / sqrt(pred) } Pur.wt1 <- nls( ~ weighted.MM1(rate, conc1, conc.1, Vm, K), data = lisTreat, start = list(Vm = 200, K = 0.1)) stopifnot(all.equal(coef(Pur.wt), coef(Pur.wt1))) ## Chambers and Hastie (1992) Statistical Models in S (p. 537): ## If the value of the right side [of formula] has an attribute called ## 'gradient' this should be a matrix with the number of rows equal ## to the length of the response and one column for each parameter. weighted.MM.grad <- function(resp, conc1, conc.1, Vm, K) { conc <- c(conc1, conc.1) K.conc <- K+conc dy.dV <- conc/K.conc dy.dK <- -Vm*dy.dV/K.conc pred <- Vm*dy.dV pred.5 <- sqrt(pred) dev <- (resp - pred) / pred.5 Ddev <- -0.5*(resp+pred)/(pred.5*pred) attr(dev, "gradient") <- Ddev * cbind(Vm = dy.dV, K = dy.dK) dev } Pur.wt.grad <- nls( ~ weighted.MM.grad(rate, conc1, conc.1, Vm, K), data = lisTreat, start = list(Vm = 200, K = 0.1)) rbind(coef(Pur.wt), coef(Pur.wt1), coef(Pur.wt.grad)) ## In this example, there seems no advantage to providing the gradient. ## In other cases, there might be. ## The two examples below show that you can fit a model to ## artificial data with noise but not to artificial data ## without noise. x <- 1:10 y <- 2*x + 3 # perfect fit ## terminates in an error, because convergence cannot be confirmed: try(nls(y ~ a + b*x, start = list(a = 0.12345, b = 0.54321))) ## adjusting the convergence test by adding 'scaleOffset' to its denominator RSS: nls(y ~ a + b*x, start = list(a = 0.12345, b = 0.54321), control = list(scaleOffset = 1, printEval=TRUE)) ## Alternatively jittering the "too exact" values, slightly: set.seed(27) yeps <- y + rnorm(length(y), sd = 0.01) # added noise nls(yeps ~ a + b*x, start = list(a = 0.12345, b = 0.54321)) ## the nls() internal cheap guess for starting values can be sufficient: x <- -(1:100)/10 y <- 100 + 10 * exp(x / 2) + rnorm(x)/10 nlmod <- nls(y ~ Const + A * exp(B * x)) plot(x,y, main = "nls(*), data, true function and fit, n=100") curve(100 + 10 * exp(x / 2), col = 4, add = TRUE) lines(x, predict(nlmod), col = 2) ## Here, requiring close convergence, must use more accurate numerical differentiation, ## as this typically gives Error: "step factor .. reduced below 'minFactor' .." \dontdiff{ try(nlm1 <- update(nlmod, control = list(tol = 1e-7))) o2 <- options(digits = 10) # more accuracy for 'trace' ## central differencing works here typically (PR#18165: not converging on *some*): ctr2 <- nls.control(nDcentral=TRUE, tol = 8e-8, # <- even smaller than above warnOnly = TRUE || # << work around; e.g. needed on some ATLAS-Lapack setups (grepl("^aarch64.*linux", R.version$platform) && grepl("^NixOS", osVersion) )) (nlm2 <- update(nlmod, control = ctr2, trace = TRUE)); options(o2) ## --> convergence tolerance 4.997e-8 (in 11 iter.) } ## The muscle dataset in MASS is from an experiment on muscle ## contraction on 21 animals. The observed variables are Strip ## (identifier of muscle), Conc (Cacl concentration) and Length ## (resulting length of muscle section). \dontdiff{ if(requireNamespace("MASS", quietly = TRUE)) withAutoprint({ ## The non linear model considered is ## Length = alpha + beta*exp(-Conc/theta) + error ## where theta is constant but alpha and beta may vary with Strip. with(MASS::muscle, table(Strip)) # 2, 3 or 4 obs per strip ## We first use the plinear algorithm to fit an overall model, ## ignoring that alpha and beta might vary with Strip. musc.1 <- nls(Length ~ cbind(1, exp(-Conc/th)), MASS::muscle, start = list(th = 1), algorithm = "plinear") summary(musc.1) ## Then we use nls' indexing feature for parameters in non-linear ## models to use the conventional algorithm to fit a model in which ## alpha and beta vary with Strip. The starting values are provided ## by the previously fitted model. ## Note that with indexed parameters, the starting values must be ## given in a list (with names): b <- coef(musc.1) musc.2 <- nls(Length ~ a[Strip] + b[Strip]*exp(-Conc/th), MASS::muscle, start = list(a = rep(b[2], 21), b = rep(b[3], 21), th = b[1])) summary(musc.2) }) } \dontshow{options(od)} } \keyword{nonlinear} \keyword{regression} \keyword{models}