Design of Experiments in Metal Cutting Tests 309
where X(x
1
, x
2
,...,x
M
) is the input variables vector, M is the number of input variables
and A(b
1
, b
2
, ..., b
M
) is the vector of coefficients.
Components of the input vector X can be independent variables and functional forms or
finite difference terms. Other non-linear reference functions, such as difference, proba-
bilistic, harmonic and logistic can also be used. The method allows finding simultaneously
the structure of model and the dependence of modeled system output on the values of
most significant inputs of the system.
GMDH, based on the self-organizing principle, requires minimum information about
the object under study. As such, all the available information about this object should
be used. The algorithm allows finding the needed additional information through the
sequential analysis of different models using the so-called external criteria. Therefore,
GMDH is a combined method: it uses the test data and sequential analysis and estimation
of the candidate models. The estimates are found using relatively small part of the test
results. The other part of these results is used to estimate the model coefficients and to
find the optimal model structure.
Although GMDH and regression analysis use the table of test data, the regression analysis
requires the prior formulation of the regression model and its complexity. This is because
the row variances used in the calculations (Section, Statistical Examination of the Result
Obtained, Eq. (5.26)) are internal criteria. A criterion is called an internal criterion if its
determination is based on the same data that is used to develop the model. The use of any
internal criterion leads to a false rule: the more complex model is more accurate. This
is because the complexity of the model is determined by the number and highest power
of its terms. As such, the greater the number of terms, the smaller the variance. GMDH
uses the external criteria. A criterion is called external if its determination is based on
new information obtained using “fresh” points of the experimental table not used in
the model development. This allows the selection of the model of optimum complexity
corresponding to the minimum of the selected external criterion.
Another significant difference between the regression analysis and GMDH is that the
former allows construction of the model only in the domain where the number of model
coefficients is less than the number of points of the design matrix because the examination
of model adequacy is possible only when f
ad
> 0, i.e. when the number of estimated
coefficients of the model (n) is less than the number of points in the design matrix (m).
GMDH allows much wider domain where, for example, the number of model coefficients
can be millions and all these are estimated using the design matrix containing only 20
rows. In this new domain, accurate and unbiased models are obtained. GMDH algorithms
utilize minimum experimental information on input. This input consists of a table having
10–20 points and the criterion of model selection. The algorithms determine the unique
model of optimal complexity by the sorting out of different models using the selected
criterion.
The essence of the self-organizing principle in GMDH is that the external criteria pass
their minimum when the complexity of the model is gradually increased. When a par-
ticular criterion is selected, the computer executing GMDH finds this minimum and the
corresponding model of optimal complexity. As such, the value of the selected criterion
referred to as the depth of minimum can be considered as an estimate of the accuracy