计算机科学
安全性令牌
模式
代表(政治)
水准点(测量)
模态(人机交互)
人工智能
语义学(计算机科学)
机器学习
多模式学习
外部数据表示
标杆管理
特征学习
程序设计语言
业务
营销
地理
法学
政治学
政治
社会学
大地测量学
计算机安全
社会科学
作者
Xianbing Zhao,Yixin Chen,Sicen Liu,Xuan Zang,Yang Xiang,Buzhou Tang
标识
DOI:10.1145/3543507.3583406
摘要
Existing methods for Multimodal Sentiment Analysis (MSA) mainly focus on integrating multimodal data effectively on limited multimodal data. Learning more informative multimodal representation often relies on large-scale labeled datasets, which are difficult and unrealistic to obtain. To learn informative multimodal representation on limited labeled datasets as more as possible, we proposed TMMDA for MSA, a new Token Mixup Multimodal Data Augmentation, which first generates new virtual modalities from the mixed token-level representation of raw modalities, and then enhances the representation of raw modalities by utilizing the representation of the generated virtual modalities. To preserve semantics during virtual modality generation, we propose a novel cross-modal token mixup strategy based on the generative adversarial network. Extensive experiments on two benchmark datasets, i.e., CMU-MOSI and CMU-MOSEI, verify the superiority of our model compared with several state-of-the-art baselines. The code is available at https://github.com/xiaobaicaihhh/TMMDA.
科研通智能强力驱动
Strongly Powered by AbleSci AI