Eng-Tips is the largest engineering community on the Internet

Intelligent Work Forums for Engineering Professionals

  • Congratulations waross on being selected by the Tek-Tips community for having the most helpful posts in the forums last week. Way to Go!

How to evaluate the different fitting function for a set of data?

Status
Not open for further replies.

elbow1

Coastal
Oct 25, 2004
11
0
0
HK
I fit a set of data (around 10 points) with several polynomial functions. All of them are fitting well with similar coefficient of correlation. I want to compare these functions and choose a best one.

Are there some other parameters or methods can be used to compare the effects of these fitting function?

Thanks in advance.
 
Replies continue below

Recommended for you

An ANOVA table will show which one of the curves is a significant fit to the data. It may show you that the signifcance of a linear fit is hardly any different to a quadratic fit, for example. In that case choose the linear fit.

corus
 
Is their any way to obtain a few more data points? Let's say that you have data points X=-15, -10.5, -.75, 0, 5.25, 8.3. Is there any way that you can get at least 4 more data point, say X=-25, -5, 2, 20? If so, take each of your "fitting functions" and predict the Y value for the new X points. This may halp you determine which function actual represents your physical system the best.
 
What are the polynomial functions?

As a general rule, higher order polynomials will fit the same curve as a simpler one, but will more bumps and wiggles that are not physically meaningful.

TTFN
 
In some cases a so called "penalty" function is introduced to reduce one's strive to include too many parameters for a better fit: the purpose of a best fit procedure is to describe your data as good as possible but with as little parametra as possible. So with each additional parameter that makes a better fit you are more far from the basic idea, why you are doing best fit and at the end you have a perfect fit but you have as many parameters as you have the data points, so what is the use of it!
In nonlinear best fit procedures a criterium (one among others) that has been widely adopted is so called Aikake information criterium (AIC)and is ussually calculated as AIC=-2*(LOG(-likelihood))+2*(number of parameters in the model).We select the model with the smallest AIC.
m777182
 
Try a variation on bootstrapping. When testing the sensitivity of a linear fit previously, I have used this to good effect.

With such a small number of data points, just remove a single point, and recalculate the fit. If the curve fit is good, then removing a single data point should not alter the fit greatly. You can try removing each of the data points to see what effect they have on the fit returned. What you are doing really is determining how sensitive your fit is to the data you have used to create the fit.

An example of this would be a linear fit on a tight cluster of points, with one outlier well outside the cluster. If you calculate a linear fit on this, you get R squared close to 1, but if you then remove that outlier, the descriptors (slope and intercept) change majorly, as does the fit quality.

This method is quite good at testing fitting graphically, and requires very little knowledge of statistics.
 
Status
Not open for further replies.
Back
Top