Title Boosting ridge for the extreme learning machine globally optimised for classification and regression problems
Authors Peralez-González C. , PÉREZ RODRÍGUEZ, JAVIER, DURAN ROSAL, ANTONIO MANUEL
External publication No
Means Scientific Reports
Scope Article
Nature Científica
JCR Quartile 1
SJR Quartile 1
JCR Impact 3.8
SJR Impact 0.9
Web https://www.scopus.com/inward/record.uri?eid=2-s2.0-85165387389&doi=10.1038%2fs41598-023-38948-3&partnerID=40&md5=d7974cae5aba4819d10dc548c4e13091
Publication date 24/04/2023
ISI 001067878900041
Scopus Id 2-s2.0-85165387389
DOI 10.1038/s41598-023-38948-3
Abstract This paper explores the boosting ridge (BR) framework in the extreme learning machine (ELM) community and presents a novel model that trains the base learners as a global ensemble. In the context of Extreme Learning Machine single-hidden-layer networks, the nodes in the hidden layer are preconfigured before training, and the optimisation is performed on the weights in the output layer. The previous implementation of the BR ensemble with ELM (BRELM) as base learners fix the nodes in the hidden layer for all the ELMs. The ensemble learning method generates different output layer coefficients by reducing the residual error of the ensemble sequentially as more base learners are added to the ensemble. As in other ensemble methodologies, base learners are selected until fulfilling ensemble criteria such as size or performance. This paper proposes a global learning method in the BR framework, where base learners are not added step by step, but all are calculated in a single step looking for ensemble performance. This method considers (i) the configurations of the hidden layer are different for each base learner, (ii) the base learners are optimised all at once, not sequentially, thus avoiding saturation, and (iii) the ensemble methodology does not have the disadvantage of working with strong classifiers. Various regression and classification benchmark datasets have been selected to compare this method with the original BRELM implementation and other state-of-the-art algorithms. Particularly, 71 datasets for classification and 52 for regression, have been considered using different metrics and analysing different characteristics of the datasets, such as the size, the number of classes or the imbalanced nature of them. Statistical tests indicate the superiority of the proposed method in both regression and classification problems in all experimental scenarios. © 2023, The Author(s).
Keywords algorithm; article; classifier; human; learning; machine learning
Universidad Loyola members

Change your preferences Manage cookies