对抗制
生成语法
生成对抗网络
人工智能
计算机科学
分布(数学)
模式识别(心理学)
机器学习
数学
深度学习
数学分析
作者
Yueqi Chen,Witold Pedrycz,Tingting Pan,Jian Wang,Jie Yang
标识
DOI:10.1002/adts.202400234
摘要
Abstract Tackling imbalanced problems encountered in real‐world applications poses a challenge at present. Oversampling is a widely useful method for imbalanced tabular data. However, most traditional oversampling methods generate samples by interpolation of minority (positive) class, failing to entirely capture the probability density distribution of the original data. In this paper, a novel oversampling method is presented based on generative adversarial network (GAN) with the originality of introducing three strategies to enhance the distribution of the positive class, called GAN‐E. The first strategy is to inject prior knowledge of positive class into the latent space of GAN, improving sample emulation. The second strategy is to inject random noise containing this prior knowledge into both original and generated positive samples to stretch the learning space of the discriminator of GAN. The third one is to use multiple GANs to learn comprehensive probability distributions of positive class based on multi‐scale data to eliminate the influence of GAN on generating aggregate samples. The experimental results and statistical tests obtained on 18 commonly used imbalanced datasets show that the proposed method comes with a better performance in terms of G‐mean, F‐measure, AUC and accuracy than 14 other rebalanced methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI