TY - JOUR
T1 - PSNet
T2 - Towards Efficient Image Restoration With Self-Attention
AU - Cui, Yuning
AU - Knoll, Alois
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2023/9/1
Y1 - 2023/9/1
N2 - Image restoration aims to recover a clean image from various degradations, e.g., haze, snow, and blur, playing an important role in robot vision, autonomous vehicles, and medical imaging. Recently, the use of Transformer has witnessed a significant improvement in multifarious image restoration tasks. However, despite a few remedies to reduce the quadratic complexity of self-attention, these approaches are still impractical for real-world applications, which need high efficiency and speed. To ameliorate this issue, we propose an efficient framework for image restoration based on self-attention. To this end, we combine the strengths of patch-based and strip-based self-attention units to improve efficiency. More specifically, we apply self-attention of different operation scales to features of different resolutions, i.e., we adopt a relatively smaller region for self-attention on high-resolution features while a larger region for low-restoration features. In addition, instead of using global self-attention in each partitioned region, we leverage a strip-based version for low complexity. To further improve efficiency, we insert our design into a U-shaped CNN network to establish our framework, dubbed PSNet. Extensive experiments demonstrate that our network receives state-of-the-art performance on five representative image restoration tasks with low computational complexity and high speed, including single-image defocus deblurring, image dehazing, image motion deblurring, image desnowing, and image denoising.
AB - Image restoration aims to recover a clean image from various degradations, e.g., haze, snow, and blur, playing an important role in robot vision, autonomous vehicles, and medical imaging. Recently, the use of Transformer has witnessed a significant improvement in multifarious image restoration tasks. However, despite a few remedies to reduce the quadratic complexity of self-attention, these approaches are still impractical for real-world applications, which need high efficiency and speed. To ameliorate this issue, we propose an efficient framework for image restoration based on self-attention. To this end, we combine the strengths of patch-based and strip-based self-attention units to improve efficiency. More specifically, we apply self-attention of different operation scales to features of different resolutions, i.e., we adopt a relatively smaller region for self-attention on high-resolution features while a larger region for low-restoration features. In addition, instead of using global self-attention in each partitioned region, we leverage a strip-based version for low complexity. To further improve efficiency, we insert our design into a U-shaped CNN network to establish our framework, dubbed PSNet. Extensive experiments demonstrate that our network receives state-of-the-art performance on five representative image restoration tasks with low computational complexity and high speed, including single-image defocus deblurring, image dehazing, image motion deblurring, image desnowing, and image denoising.
KW - Deep learning for visual perception
KW - representation learning
KW - visual learning
UR - http://www.scopus.com/inward/record.url?scp=85166771646&partnerID=8YFLogxK
U2 - 10.1109/LRA.2023.3300254
DO - 10.1109/LRA.2023.3300254
M3 - Article
AN - SCOPUS:85166771646
SN - 2377-3766
VL - 8
SP - 5735
EP - 5742
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
IS - 9
ER -