← Volver atrás
Publicaciones

Oversampling the Minority Class in the Feature Space

Autores

PÉREZ ORTIZ, MARÍA, Gutierrez P.A. , Tino P. , Hervas-Martinez C.

Publicación externa

No

Medio

IEEE Trans. Neural Netw. Learn. Syst.

Alcance

Article

Naturaleza

Científica

Cuartil JCR

Cuartil SJR

Impacto JCR

6.108

Impacto SJR

2.564

Fecha de publicacion

01/01/2016

ISI

000382175500012

Scopus Id

2-s2.0-84940706270

Abstract

The imbalanced nature of some real-world data is one of the current challenges for machine learning researchers. One common approach oversamples the minority class through convex combination of its patterns. We explore the general idea of synthetic oversampling in the feature space induced by a kernel function (as opposed to input space). If the kernel function matches the underlying problem, the classes will be linearly separable and synthetically generated patterns will lie on the minority class region. Since the feature space is not directly accessible, we use the empirical feature space (EFS) (a Euclidean space isomorphic to the feature space) for oversampling purposes. The proposed method is framed in the context of support vector machines, where the imbalanced data sets can pose a serious hindrance. The idea is investigated in three scenarios: 1) oversampling in the full and reduced-rank EFSs; 2) a kernel learning technique maximizing the data class separation to study the influence of the feature space structure (implicitly defined by the kernel function); and 3) a unified framework for preferential oversampling that spans some of the previous approaches in the literature. We support our investigation with extensive experiments over 50 imbalanced data sets. © 2012 IEEE.

Palabras clave

Artificial intelligence; Convex combinations; Empirical feature space; Euclidean spaces; Imbalanced Data-sets; Kernel function; Kernel learning; Linearly separable; Unified framework; Learning systems