HAAT: hybrid attention aggregation transformer for image super-resolution
变压器
计算机科学
计算机视觉
人工智能
模式识别(心理学)
电气工程
电压
工程类
作者
S. Anil Lal,Tsun-Hin Cheung,Kin Yip Fung,Kaixin Xue,Kin‐Man Lam
标识
DOI:10.1117/12.3058003
摘要
In the research area of image super-resolution, Swin-transformer-based models are favored for their global spatial modeling and shifting window attention mechanism. However, existing methods often limit self-attention to nonoverlapping windows to cut costs and ignore the useful information that exists across channels. To address this issue, this paper introduces a novel model, the Hybrid Attention Aggregation Transformer (HAAT), designed to better leverage feature information. HAAT is constructed by integrating Swin-Dense-Residual-Connected Blocks (SDRCB) with Hybrid Grid Attention Blocks (HGAB). SDRCB expands the receptive field while maintaining a streamlined architecture, resulting in enhanced performance. HGAB incorporates channel attention, sparse attention, and window attention to improve nonlocal feature fusion and achieve more visually compelling results. Experimental evaluations demonstrate that HAAT surpasses state-of-the-art methods on benchmark datasets.