已入深夜,您辛苦了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!祝你早点完成任务,早点休息,好梦!

CrossFormer: Cross-modal Representation Learning via Heterogeneous Graph Transformer

计算机科学 情态动词 变压器 图形 代表(政治) 理论计算机科学 人工智能 电压 化学 物理 量子力学 政治 政治学 高分子化学 法学
作者
Xiao Liang,Erkun Yang,Cheng Deng,Yanhua Yang
出处
期刊:ACM Transactions on Multimedia Computing, Communications, and Applications [Association for Computing Machinery]
被引量:4
标识
DOI:10.1145/3688801
摘要

Transformers have been recognized as powerful tools for various cross-modal tasks due to their superior ability to perform representation learning through self-attention. Existing transformer-based cross-modal models can be categorized into single-stream and dual-stream ones. By performing fine-grained interaction with self-attention on the cross-modal concatenated features, the former can simultaneously learn intra- and inter-modal correlations. However, this simple concatenation treats the inputs of different modalities equally; as a result, the heterogeneous differences between modalities are ignored, leading to a modality gap. The latter process the inputs of different modalities separately, then perform cross-modal interaction on the subsequently fused networks, resulting in a failure to integrate the fine-grained correlations of both intra- and inter-modality in a uniform module. To this end, we propose an effective heterogeneous graph transformer for dual-stream cross-modal representation learning, named CrossFormer, which constructs a heterogeneous graph as a bridge to achieve fine-grained intra- and inter-modal interaction on a dual-stream network. Specifically, we first represent multi-modal data with a heterogeneous graph, then develop a dual-positional encoding strategy that enables the heterogeneous graph to obtain the relative positional information. Finally, a dual-stream self-attention is performed on the heterogeneous graph, bridging the gap between modalities and effectively capturing fine-grained intra- and inter-modal interactions simultaneously. Extensive experiments on various cross-modal tasks demonstrate the superiority of our method.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
zjl1112发布了新的文献求助10
1秒前
可莉完成签到 ,获得积分10
1秒前
3秒前
领导范儿应助csq69采纳,获得10
3秒前
常绕凌淑完成签到,获得积分10
4秒前
5秒前
yff完成签到,获得积分10
5秒前
11发布了新的文献求助10
5秒前
5秒前
8秒前
zjl1112完成签到,获得积分10
9秒前
orixero应助xcf采纳,获得10
10秒前
AU发布了新的文献求助10
10秒前
zeizei发布了新的文献求助10
11秒前
慕青应助叁叁采纳,获得10
12秒前
rtkndg完成签到 ,获得积分20
12秒前
敏感初露发布了新的文献求助10
12秒前
gao0505完成签到,获得积分10
12秒前
单纯向雪完成签到 ,获得积分10
12秒前
Orange应助csq69采纳,获得10
13秒前
haha发布了新的文献求助10
13秒前
满意夏岚完成签到,获得积分20
13秒前
xueshufengbujue完成签到,获得积分10
16秒前
pancover完成签到,获得积分20
16秒前
17秒前
Yasong完成签到 ,获得积分10
18秒前
18秒前
18秒前
背后的不惜完成签到 ,获得积分10
19秒前
竹林完成签到,获得积分20
20秒前
21秒前
21秒前
22秒前
刘很红发布了新的文献求助10
22秒前
25秒前
慕青应助阿洁采纳,获得30
26秒前
隐形的蚂蚁完成签到 ,获得积分10
30秒前
Dr_JennyZ发布了新的文献求助20
30秒前
AC咪咪发布了新的文献求助50
31秒前
深情安青应助凌香芦采纳,获得10
31秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Salmon nasal cartilage-derived proteoglycan complexes influence the gut microbiota and bacterial metabolites in mice 2000
The Composition and Relative Chronology of Dynasties 16 and 17 in Egypt 1500
Picture this! Including first nations fiction picture books in school library collections 1500
SMITHS Ti-6Al-2Sn-4Zr-2Mo-Si: Ti-6Al-2Sn-4Zr-2Mo-Si Alloy 850
Signals, Systems, and Signal Processing 610
Learning manta ray foraging optimisation based on external force for parameters identification of photovoltaic cell and module 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6376042
求助须知:如何正确求助?哪些是违规求助? 8189329
关于积分的说明 17293420
捐赠科研通 5429948
什么是DOI,文献DOI怎么找? 2872782
邀请新用户注册赠送积分活动 1849306
关于科研通互助平台的介绍 1694974