首页    期刊浏览 2024年11月28日 星期四
登录注册

文章基本信息

  • 标题:The Effect of Different Dimensionality Reduction Techniques on Machine Learning Overfitting Problem
  • 本地全文:下载
  • 作者:Mustafa Abdul Salam ; Ahmad Taher Azar ; Mustafa Samy Elgendy
  • 期刊名称:International Journal of Advanced Computer Science and Applications(IJACSA)
  • 印刷版ISSN:2158-107X
  • 电子版ISSN:2156-5570
  • 出版年度:2021
  • 卷号:12
  • 期号:4
  • 页码:641-655
  • DOI:10.14569/IJACSA.2021.0120480
  • 出版社:Science and Information Society (SAI)
  • 摘要:In most conditions, it is a problematic mission for a machine-learning model with a data record, which has various attributes, to be trained. There is always a proportional relationship between the increase of model features and the arrival to the overfitting of the susceptible model. That observation occurred since not all the characteristics are always important. For example, some features could only cause the data to be noisier. Dimensionality reduction techniques are used to overcome this matter. This paper presents a detailed comparative study of nine dimensionality reduction methods. These methods are missing-values ratio, low variance filter, high-correlation filter, random forest, principal component analysis, linear discriminant analysis, backward feature elimination, forward feature construction, and rough set theory. The effects of used methods on both training and testing performance were compared with two different datasets and applied to three different models. These models are, Artificial Neural Network (ANN), Support Vector Machine (SVM) and Random Forest classifier (RFC). The results proved that the RFC model was able to achieve the dimensionality reduction via limiting the overfitting crisis. The introduced RFC model showed a general progress in both accuracy and efficiency against compared approaches. The results revealed that dimensionality reduction could minimize the overfitting process while holding the performance so near to or better than the original one.
  • 关键词:Dimensionality reduction; feature subset selection; rough set; overfitting; underfitting; machine learning
国家哲学社会科学文献中心版权所有