# 12.1. Fitting Overview¶

Modeling and fitting of experimental data is a key need for the analysis of most scientific data. There is an extensive literature on these topics, with a wealth written on both the theoretical and practical aspects of modeling data. One of the more common and general approaches is to use a least-squares analysis, in which a parameterized model is adjusted until it matches experimental data in the sense that the sum of squares of the difference between data and model is as small as possible. Mathematically, this is expressed as

$S = \sum_{i=1}^{N} \big[{y_i - f(x_i, \vec{\beta}) } \big]^2$

where the experimental data is expressed as $$y(x)$$ that is discretely sampled at $$N$$ points, $$f(x, \vec\beta)$$ is a model function of some dependent data $$x$$ and $$\vec\beta$$, a set of parameters in the model. As a simple example, a linear model of data would be written as $$f = \beta_0 + \beta_1{x}$$. Of course, models can be arbitrary complex. There is good statistical justification for using the least-squares approach, and many existing tools for helping to find the minimal values of $$S$$. These justifcations are not without criticism or caveats, but we’ll leave those topics aside for now.

It is common to include a multiplicative factor to each component in the least-squares equation above, so that the different samples (or data points) might be given more or less weight. Again, there are several approaches to this, with one of the most common approaches to weight by the inverse of the estimated uncertainty in the data value. This then what is generally called the chi-square goodness-of-fit parameter

$\chi^2 = \sum_{i=1}^{N} \big[\frac{y_i - f(x_i, \vec{\beta})}{\epsilon_i} \big]^2$

Here, $$\epsilon_i$$ represents the uncertainty in the value of $$y_i$$. As mentioned, the model describing $$f(x, \vec{\beta})$$ can be fairly complex.

There is an extensive literature for the case in which the model function $$f(x, \vec{\beta})$$ depends linearly on its parameters $$\vec{\beta}$$ (but not necessarily linearly on a dependent variable – a quadratic function $$\beta_0 + \beta_1 x + \beta_2 x^2$$ is linear in this sense). Of course, model function need not be linear in its parameters, and the minimization is generally referred to a ‘’non-linear least-squares optimization’’ in the literature. All the discussion here will assume that the models can be non-linear.

It is convenient to define the residual array $$r$$ with $$N$$ elements:

$r_i = \frac{y_i - f(x_i, \vec\beta)}{\epsilon_i}$

so that the sum to be minimized is a simple sum of this function, $$s = \sum_i^{N} r_i^2$$. The fitting process can then be made very general with a few key components required. Specifically, for Larch, the requirements are

1. A set of Parameters, $${\vec{\beta}}$$, that are used in the model, and are to be adjusted to find the least-square value of the sum of squares of the residual. These must be parameters (discussed below) that are held in a single parameter group. This can either be simple Larch group or a specialized ParameterGroup (see param_group(): creating a Parameter Group for fitting constraints). With both options, the group can contain non-parameter values as well as parameters.

2. An objective function to calculate the residual array. This will be a Larch function that takes the parameter group described above as its first argument, and an unlimited set of optional arguments. This function should return the residual array, $$r$$ that will be minimized in the least-squares sense. The arrays for the data can be passed into this function either by the optional arguments or by putting these data in the parameter group.

After the fit has completed, several statistical results describing the fit quality and the values and uncertainties found for the parameters will be stored the result.

Because the objective function will be called by the fitting process, it needs to follow fairly strict guidelines in its inputs and outputs. Specifically, the first argument to the function must be a Larch group containing all the Parameters in the model. Furthermore, the return value of the objective function must be the fit residual array to be minimized in the least-squares sense.

We’ll jump in with a simple example fit to a line, with this script:

# create mock data for a line
dat = group(x = linspace(0, 10, 51))
dat.y = 1.0 + 2.5 * dat.x + random.normal(size=51, scale=1)

# create a group of fit parameters
params = param_group(off = guess(0), slope = guess(0))

init = params.off + params.slope * dat.x

# define objective function for fit residual
def fitresid(p, data):
return data.y - (p.off + p.slope * data.x)
enddef

# perform fit
result = minimize(fitresid, params, args=(dat,))

# create final model using best-fit values for the parameters.
final = params.off + params.slope * dat.x


Here params is the parameter group, with both params.off and params.slope as Parameters that will be adjusted in the fit. fitresid is the objective function that calculates and returns the residual array (data - model) from the values of the Parameters. The minimize() function does the actual fit, and will call the objective function many times with different (and generally improving) values for the parameters. Of course, there are faster and more statistically sound methods for determining a linear trend in a set of data, but the point of this example is to illustrate the mechanisms for doing more complex, non-linear modeling of data.