|
Software: |
The method of LEAST SQUARES is used. Linear Regression To
illustrate, if z = A0 + A1*x
+ A2*y is to
be fitted to N data points, the normal equations are A0*N + A1*åx + A2*åy = åz A0*åx + A1*åx*x + A2*åx*y = åx*z A0*åy + A1*åy*x + A2*åy*y = åy*z
Solving the above system we get the values of the coefficients A0, A1 and A2. Obviously, it may be generalized to any number of variables or parameters. The correlation coefficient is computed with
å
[festimated - fmean]2 The standard error of estimate is calculated using
å
[f - festimated ]2 Summations are over all N data points, f is the dependent variable, and the festimated is found using the regression results. Curvilinear Regression If we wish to fit the curve F(W) = A0*G(x,y,z) + A1*H(x,y,z) + A2*J(x,y,z) to N data points then the normal equations are A0*å G*G + A1*å G*H + A2*å G*J = å G*F A0*å H*G + A1*å H*H + A2*å H*J = å H*F A0*å J*G + A1*å J*H + A2*å J*J = å J*F where the summations are over all N data points. Solving the above system we get the values of the coefficients A0, A1 and A2. Obviously, it may be generalized to any number of variables or parameters. The correlation coefficient and the standard error of estimate are calculated as in LINEAR REGRESSION. General Regression To illustrate the general method, let's say we want to fit a curve with three parameters, A, B and C, to N data points, that is, the curve F(A,B,C,x) where x standS for any number of variables. If A0, B0 and C0 are sufficiently close approximations to A, B and C, then we can use Taylor's theorem to put F(A,B,C,x) ÷ F(A0,B0,C0,x) + (A-A0)*Fa + (B-B0)*Fb + (C-C0)*Fc where we ignored terms of order higher than the first and where
¶F
¶F
¶F Now, let's call the residuals R = F(A0,B0,C0,x) - F(A,B,C,x) If we
minimize the residuals using the method of least squares, we obtain the
normal equations (A-A0)*å Fb*Fa + (B-B0)*å Fb*Fb + (C-C0)*å Fb*Fc = - å Fb*R (A-A0)*å Fc*Fa + (B-B0)*å Fc*Fb + (C-C0)*å Fc*Fc = - å Fc*R where Fa, Fb and Fc are as defined before and the summations are over all N data points. If we solve this last set of linear equations for (A-A0), (B-B0) and (C-C0), we obtain corrections on the initial approximation, that is, if the solutions to the system are DA, DB and DC, where A-A0=DA ; B-B0=DB ; C-C0=DC then a closer approximation to the solution is A1=A0+DA ; B1=B0+DB ; C1=C0+DC We may now use A1, B1 and C1 to again set a system of linear equations and solve it to obtain A2, B2 and C2. This is done until all parameters in the current approximation in the sequence differ from those of the previous approximation by less than their specified error and the change in R² is less than its specified error. Of course, the method may be generalized to any number of variables or parameters. The correlation coefficient and the standard error of estimate are calculated as in LINEAR REGRESSION. |
|
Copyright © 2001-2010 Numerical Mathematics. All rights reserved.
|