Non-Linear Curve Fits
INTRODUCTION
Frequently we have a table of values as a function of a single variable which may have measurement noise or other inaccuracies. These would be of the form
What we want to do then is to find the “m” parameters “λ” of some function f(x; λ) which is “non-linear” in the parameters and provides the best fit to the tabulated data.
NON-LINEAR FITS
Given some function, then at each data point we can write
for the “m” parameters
and where “” represents the uncertainty or measurement error as a consequence of an imperfect initial estimate of the parameters. We can then approximate the errors as a linear combination of corrections to our parameter set as
where which could be used to iteratively update the original set as
Writing out the entire matrix we get
or more concisely
But this is an over determined set of equations and not in form allowing for a unique determination of the parameter corrections, . So we multiply each side by the transpose of the matrix “A” which switches rows and columns as
So that
where
and for each element
And also by recalling
then
or finally
which allows for a unique solution for values.
GAUSSIAN FUNCTION
The well known Gaussian distribution (also called the “Normal” distribution or Bell curve) has the general functional form
where the three parameters are given by
In this parameter set, A is area under the curve of G(x). The arithmetic average or mean is . And the last parameter σ is the standard distribution. These are further described in the Appendix: Gaussian Parameters.
We can write the derivatives of the Gaussian with respect to its three parameters as follows
The iterative corrections to the initial guess for the three parameters can be expressed in the equation
or in detail, this expands to the following
For the measurements of
then the individual matrix coefficients are
And defining
then the matrix elements are
We could solve the matrix equation using Gauss-Jordan but for only three parameters it is easier to write the solutions directly as
And this allows us to iteratively update the initial guesses for the parameters
APPENDIX: GAUSSIAN PARAMETERS
For a random variable, X, perhaps with a Gaussian probability distribution, the results of “n” independent measurements can be represented as
The arithmetic average or mean of this series of “n” values is simply given by
and the variance (or square of the standard distribution) for a large number of “n” trials can be written
These equations are exact and work well for a small number of samples. But as the number of samples increase, it is easier to create a histogram.
Basically the data is grouped into a set of “m” bins where “yj” represents the number of values which lie within some small range, “Δx”, around a midpoint of “xj”. Each small range represents a separate bin and can be drawn as
As before the expression for the mean is
and the variance can be written as
As the number of discrete bins increases and the width “Δx” of each gets smaller, then the bins come to resemble a continuous curve as follows:
For a continuous curve the area under the Gaussian is given by
The mean value can then be written as
And the variance is given by
APPENDIX: INITIAL ESTIMATE
For a discrete set of data points
then if we first define
then the arithmetic average or mean value of the distribution is given by
and the variance or square of the standard deviation is given by
But since we are linear in A, we can write
where
then the total error associated with an initial guess for A is
and we can minimize A in a least squares sense by setting the derivative to zero
or