Título Ordinal Neural Networks Without Iterative Tuning
Autores FERNÁNDEZ NAVARRO, FRANCISCO DE ASÍS, Riccardi, Annalisa , Carloni, Sante
Publicación externa Si
Medio IEEE Trans. Neural Netw. Learn. Syst.
Alcance Article
Naturaleza Científica
Cuartil JCR 1
Cuartil SJR 1
Impacto JCR 4.29100
Impacto SJR 2.41000
Fecha de publicacion 01/11/2014
ISI 000344089300011
DOI 10.1109/TNNLS.2014.2304976
Abstract Ordinal regression (OR) is an important branch of supervised learning in between the multiclass classification and regression. In this paper, the traditional classification scheme of neural network is adapted to learn ordinal ranks. The model proposed imposes monotonicity constraints on the weights connecting the hidden layer with the output layer. To do so, the weights are transcribed using padding variables. This reformulation leads to the so-called inequality constrained least squares (ICLS) problem. Its numerical solution can be obtained by several iterative methods, for example, trust region or line search algorithms. In this proposal, the optimum is determined analytically according to the closed-form solution of the ICLS problem estimated from the Karush-Kuhn-Tucker conditions. Furthermore, following the guidelines of the extreme learning machine framework, the weights connecting the input and the hidden layers are randomly generated, so the final model estimates all its parameters without iterative tuning. The model proposed achieves competitive performance compared with the state-of-the-art neural networks methods for OR.
Palabras clave Extreme learning machine (ELM); neural networks; ordinal regression (OR)
Miembros de la Universidad Loyola

Change your preferences Gestionar cookies