Title |
A multi-class classification model with parametrized target outputs for randomized-based feedforward neural networks |
Authors |
DURAN ROSAL, ANTONIO MANUEL, Durán-Fernández A. , FERNÁNDEZ NAVARRO, FRANCISCO DE ASÍS, CARBONERO RUZ, MARIANO |
External publication |
No |
Means |
APPLIED SOFT COMPUTING |
Scope |
Article |
Nature |
Científica |
JCR Quartile |
1 |
SJR Quartile |
1 |
JCR Impact |
8.7 |
SJR Impact |
1.882 |
Web |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85144047270&doi=10.1016%2fj.asoc.2022.109914&partnerID=40&md5=f163329e3f5dad0f84034605c851cdab |
Publication date |
16/12/2022 |
ISI |
000903989600007 |
Scopus Id |
2-s2.0-85144047270 |
DOI |
10.1016/j.asoc.2022.109914 |
Abstract |
Randomized-based Feedforward Neural Networks approach regression and classification (binary and multi-class) problems by minimizing the same optimization problem. Specifically, the model parameters are determined through the ridge regression estimator of the patterns projected in the hidden layer space (randomly generated in its neural network version) for models without direct links and the patterns projected in the hidden layer space along with the original input data for models with direct links. The targets are encoded for the multi-class classification problem according to the 1-of-J encoding (J the number of classes), which implies that the model parameters are estimated to project all the patterns belonging to its corresponding class to one and the remaining to zero. This approach has several drawbacks, which motivated us to propose an alternative optimization model for the framework. In the proposed optimization model, model parameters are estimated for each class so that their patterns are projected to a reference point (also optimized during the process), whereas the remaining patterns (not belonging to that class) are projected as far away as possible from the reference point. The final problem is finally presented as a generalized eigenvalue problem. Four models are then presented: the neural network version of the algorithm and its corresponding kernel version for the neural networks models with and without direct links. In addition, the optimization model has also been implemented in randomization-based multi-layer or deep neural networks. The empirical results obtained by the proposed models were compared to those reported by state-of-the-art models in the correct classification rate and a separability index (which measures the degree of separability in projection terms per class of the patterns belonging to the class of the others). The proposed methods show very competitive performance in the separability index and prediction accuracy compared to the neural networks version of the comparison methods (with and without direct links). Remarkably, the model provides significantly superior performance in deep models with direct links compared to its deep model counterpart. © 2022 The Authors |
Keywords |
Classification (of information); Classifiers; Eigenvalues and eigenfunctions; Learning systems; Multilayer neural networks; Optimization; Parameter estimation; Regression analysis; Direct links; Extreme learning machine; Functional link neural network; Generalized eigenvalue problems; Kernel-methods; Learning machines; Modeling parameters; Random vector functional link neural network; Random vectors; Randomized-based feedforward neural network; Deep neural networks |
Universidad Loyola members |
|