Support vector regression is an improvement for principle component analysis

Authors

Mohamed, A. M., Abdel Latif, S. H.
Faculty of Graduate Studies for Statistical Research, Cairo University, Cairo, Egypt.

Alwan, A. S.
College of Administration & Economics, Sulaimani University, Sulaimani, Iraq.

Abstract

The principle component analysis is used more frequently as a variables reduction technique. And recently, an evolving group of studies makes use of machine learning regression algorithms to improve the estimation of empirical models. One of the most frequently used machines learning regression models is support vector regression with various kernel functions. However, an ensemble of support vector regression and principal component analysis is also possible. So, this paper aims to investigate the competence of support vector regression techniques after performing principal component analysis to explore the possibility of reducing data and having more accurate estimations. Some new proposals are introduced and the behavior of two different models ๐œ€๐œ€-SVR and ๐‘ฃ๐‘ฃ-SVR are compared through an extensive simulation study under four different kernel functions; linear, radial, polynomial, and sigmoid kernel functions, with different sample sizes, ranges from small, moderate to large. The models are compared with their counterparts in terms of coefficient of determination (๐‘…๐‘…2 ) and root mean squared error (RMSE). The comparative results show that applying SVR after PCA models improve the results in terms of SV numbers between 30% and 60% on average and it can be applied with real data. In addition, the linear kernel function gave the best values rather than other kernel functions and the sigmoid kernel gave the worst values. Under ๐œ€๐œ€-SVR the results improved which did not happen with ๐‘ฃ๐‘ฃ-SVR. It is also drawn that, RMSE values decreased with increasing sample size.