Sparse self-attention transformer for image inpainting

修补 变压器 计算机科学 人工智能 预处理器 Softmax函数 计算复杂性理论 模式识别(心理学) 机器学习 算法 深度学习 图像(数学) 工程类 电压 电气工程
作者
Wenli Huang,Ye Deng,S. Hui,Yang Wu,Sanping Zhou,Jinjun Wang
出处
期刊:Pattern Recognition [Elsevier]
卷期号:145: 109897-109897 被引量:19
标识
DOI:10.1016/j.patcog.2023.109897
摘要

Learning-based image inpainting methods have made remarkable progress in recent years. Nevertheless, these methods still suffer from issues such as blurring, artifacts, and inconsistent contents. The use of vanilla convolution kernels, which have limited perceptual fields and spatially invariant kernel coefficients, is one of the main causes for these problems. In contrast, the multi-headed attention in the transformer can effectively model non-local relations among input features by generating adaptive attention scores. Therfore, this paper explores the feasibility of employing the transformer model for the image inpainting task. However, the multi-headed attention transformer blocks pose a significant challenge due to their overwhelming computational cost. To address this issue, we propose a novel U-Net style transformer-based network for the inpainting task, called the sparse self-attention transformer (Spa-former). The Spa-former retains the long-range modeling capacity of transformer blocks while reducing the computational burden. It incorporates a new channel attention approximation algorithm that reduces attention calculation to linear complexity. Additionally, it replaces the canonical softmax function with the ReLU function to generate a sparse attention map that effectively excludes irrelevant features. As a result, the Spa-former achieves effective long-range feature modeling with fewer parameters and lower computational resources. Our empirical results on challenging benchmarks demonstrate the superior performance of our proposed Spa-former over state-of-the-art approaches.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
1秒前
酷波er应助咚咚咚采纳,获得10
1秒前
爆米花应助cdjyoona采纳,获得10
1秒前
tramp应助思晗采纳,获得10
2秒前
可爱的函函应助YXH采纳,获得10
2秒前
刘尹发布了新的文献求助10
2秒前
大模型应助Water103采纳,获得30
2秒前
ZM发布了新的文献求助20
4秒前
难两全发布了新的文献求助10
4秒前
4秒前
安醉香发布了新的文献求助30
5秒前
大个应助戴戴采纳,获得10
5秒前
脑洞疼应助Tonald Yang采纳,获得10
5秒前
5秒前
七彩完成签到,获得积分10
6秒前
6秒前
善学以致用应助15847348169采纳,获得10
6秒前
科研通AI2S应助Wiesen采纳,获得10
6秒前
7秒前
小蒋发布了新的文献求助10
7秒前
1111111发布了新的文献求助10
8秒前
8秒前
谁也认不出我略略略完成签到,获得积分10
9秒前
Ava应助科研通管家采纳,获得10
9秒前
共享精神应助科研通管家采纳,获得10
9秒前
Lucas应助科研通管家采纳,获得10
9秒前
orixero应助科研通管家采纳,获得10
9秒前
9秒前
SciGPT应助科研通管家采纳,获得10
9秒前
9秒前
科研通AI2S应助科研通管家采纳,获得10
9秒前
9秒前
小蘑菇应助科研通管家采纳,获得10
10秒前
汉堡包应助科研通管家采纳,获得10
10秒前
乐乐应助科研通管家采纳,获得10
10秒前
MrT完成签到 ,获得积分10
10秒前
吴彦祖应助Wednesday Chong采纳,获得50
10秒前
Be-a rogue发布了新的文献求助10
10秒前
陶醉听芹发布了新的文献求助30
11秒前
善学以致用应助狗大王采纳,获得20
12秒前
高分求助中
Exploring Mitochondrial Autophagy Dysregulation in Osteosarcoma: Its Implications for Prognosis and Targeted Therapy 2000
Impact of Mitophagy-Related Genes on the Diagnosis and Development of Esophageal Squamous Cell Carcinoma via Single-Cell RNA-seq Analysis and Machine Learning Algorithms 2000
Raising Girls With ADHD: Secrets for Parenting Healthy, Happy Daughters 1000
QMS18Ed2 | process management. 2nd ed 600
LNG as a marine fuel—Safety and Operational Guidelines - Bunkering 560
The Intuitive Guide to Fourier Analysis and Spectral Estimation with MATLAB 500
晶体非线性光学:带有 SNLO 示例(第二版) 500
热门求助领域 (近24小时)
化学 医学 材料科学 生物 工程类 有机化学 生物化学 物理 内科学 纳米技术 计算机科学 化学工程 复合材料 基因 遗传学 物理化学 催化作用 免疫学 细胞生物学 电极
热门帖子
关注 科研通微信公众号,转发送积分 2948663
求助须知:如何正确求助?哪些是违规求助? 2609502
关于积分的说明 7028669
捐赠科研通 2249353
什么是DOI,文献DOI怎么找? 1193546
版权声明 590604
科研通“疑难数据库(出版商)”最低求助积分说明 583965