Bayes regularization in the selection of weight coefficients in the predictor ensembles
https://doi.org/10.15514/ISPRAS-2019-31(4)-7
Abstract
The supervised learning problem is discussed in the article: it is necessary to restore the dependence that maps a vector set into a scalar based on a finite set of examples of such a mapping - a training sample. This problem belongs to the class of inverse problems, and, like most inverse problems, is mathematically incorrect. This is expressed in the fact that if you construct the solution using the least squares method according to the points of the training sample, you may encounter retraining – a situation where the model describes the training set well, but gives a big error on the test one. We apply the approach when a solution is sought in the form of an ensemble of predictive models. Ensembles are built using the bagging method. Perceptrons and decision trees are considered as basic learning models. The final decision is obtained by weighted voting of predictors. Weights are selected by minimizing model errors in the training set. To avoid over-fitting in the selection of weights, Bayesian regularization of the solution is applied. In order to choose regularization parameters, it is proposed to use the method of orthogonalized basic functions, which allows obtaining their optimal values without using expensive iterative procedures.
About the Author
Anton Sergeevich NuzhnyRussian Federation
Candidate of Physics and Mathematics, Research Fellow
References
1. H.Zhu, R.Rohwer. No free lunch for cross-validation. Neural Computation, vol. 8, issue 7, 1996, pp. 1421-1426.
2. David J. C. MacKay. Bayesian Interpolation. Neural Computation, vol. 4, issue 3, 1992, pp. 415-447.
3. Breiman L. Bagging predictors. Machine Learning, vol. 24, no. 2, 1996, pp. 123–140.
4. А.С. Нужный, С.А. Шумский. Регуляризация Байеса в задаче аппроксимации функции многих переменных. Математическое моделирование, 2003, том 15, №9, стр. 55-63 / A.S. Nuzhny, S.A. Shumsky. The Вayes regularization in the problem of function of many variables approximation. Matematical Modeling, vol. 15, no. 9, pp. 2003, pp. 55–63 (in Russian).
5. A.S. Nuzhny. Bayesian regularization in the problem of point-by-point function approximation using orthogonalized basis. Mathematical Models and Computer Simulations, vol. 4, issue 2, 2012, pp. 203-209.
6. Simon Haykin. Neural Networks: A Comprehensive Foundation (2nd Edition). Prentice Hall, 1998, 842 p.
7. Breiman L., Friedman J.H., Stone C.J. , Olshen R.A. Classification and regression trees. Chapman and Hall/CRC, 1984, 368 p.
8. Breiman Leo. Random Forests. Machine Learning, vol. 45, no. 1, (2001, pp. 5—32.
9. Harrison D. and Rubinfeld D.L. Hedonic prices and the demand for clean air. Journal of Environmental Economics and Management, vol. 5, no. 1, 1978, pp. 81-102.
Review
For citations:
Nuzhny A.S. Bayes regularization in the selection of weight coefficients in the predictor ensembles. Proceedings of the Institute for System Programming of the RAS (Proceedings of ISP RAS). 2019;31(4):113-120. (In Russ.) https://doi.org/10.15514/ISPRAS-2019-31(4)-7