You may have noticed something, however: Starting with any set of data, we can always draw a curve in fact, lots of curves that "connect the dots" and fit the data perfectly. The sum of the squares of the residuals will be zero. So what's all the fuss?
There are at least two problems with simply "connecting the dots." One is that a curve that fits every bump in the data may have a quite complicated equation. The second is that all of those bumps, as we have noted, may be due to noise or experimental error, and not genuine trends in the data. What we want is an equation with a simple description that follows the data's general trend while ignoring the noisy bumps. This takes just a little fuss.
But not much. All we have to do is agree on what a "simple description" will be. This means picking a functional form from among some simple families of functions:
|
How the method of least squares chooses the "best" a, b, c, and d, from among all of the possible choices, alas, is a matter for calculus. For now, we will trust the computer to carry out the method accurately and efficiently for us.
This doesn't absolve us from the responsibility of first picking a reasonable family to fit the data's trend. The method of least squares will only optimize a curve within a given family.
This general procedure for finding an equation to describe a set of data points is called parametric fitting. It is a particular kind of regression analysis. (General regression analyses vary not only the parameters, but the functional family.) Here's a review of the steps:
Back to Contents |