Skip to main content

SATSal: A Multi-level Self-Attention Based Architecture for Visual Saliency Prediction.

Tliba, M., Kerkouri, M.A., Ghariba, B., Chetouani, A., Coltekin, A., Shehata, M. and Bruno, A., 2022. SATSal: A Multi-level Self-Attention Based Architecture for Visual Saliency Prediction. IEEE Access, 10, 20701-20713.

Full text available as:

[img]
Preview
PDF (OPEN ACCESS ARTICLE)
SATSal_A_Multi-Level_Self-Attention_Based_Architecture_for_Visual_Saliency_Prediction (1).pdf - Published Version
Available under License Creative Commons Attribution.

1MB

DOI: 10.1109/ACCESS.2022.3152189

Abstract

Human visual Attention modelling is a persistent interdisciplinary research challenge, gaining new interest in recent years mainly due to the latest developments in deep learning. That is particularly evident in saliency benchmarks. Novel deep learning-based visual saliency models show promising results in capturing high-level (top-down) human visual attention processes. Therefore, they strongly differ from the earlier approaches, mainly characterised by low-level (bottom-up) visual features. These developments account for innate human selectivity mechanisms that are reliant on both high- and low-level factors. Moreover, the two factors interact with each other. Motivated by the importance of these interactions, in this project, we tackle visual saliency modelling holistically, examining if we could consider both high- and low-level features that govern human attention. Specifically, we propose a novel method SAtSal (Self-Attention Saliency). SAtSal leverages both high and low-level features using a multilevel merging of skip connections during the decoding stage. Consequently, we incorporate convolutional self-attention modules on skip connection from the encoder to the decoder network to properly integrate the valuable signals from multilevel spatial features. Thus, the self-attention modules learn to filter out the latent representation of the salient regions from the other irrelevant information in an embedded and joint manner with the main encoder-decoder model backbone. Finally, we evaluate SAtSal against various existing solutions to validate our approach, using the well-known standard saliency benchmark MIT300. To further examine SAtSal’s robustness on other image types, we also evaluate it on the Le-Meur saliency painting benchmark.

Item Type:Article
ISSN:2169-3536
Uncontrolled Keywords:Eye Movements; Low and High vision; Saliency Prediction; Self-attention; Visual Attention
Group:Faculty of Science & Technology
ID Code:36698
Deposited By: Symplectic RT2
Deposited On:01 Mar 2022 15:58
Last Modified:18 May 2022 15:47

Downloads

Downloads per month over past year

More statistics for this item...
Repository Staff Only -