Praveen, Arockia, Noorwali, Abdulfattah, Samiayya, Duraimurugan, Khan, Mohammad Zubair, Vincent, PMDR, Bashir, Ali Kashif ORCID: https://orcid.org/0000-0001-7595-2522 and Alagupandi, Vinoth (2021) ResMem-Net: memory based deep CNN for image memorability estimation. PeerJ Computer Science, 7. e767-e767. ISSN 2376-5992
|
Published Version
Available under License Creative Commons Attribution. Download (759kB) | Preview |
Abstract
Image memorability is a very hard problem in image processing due to its subjective nature. But due to the introduction of Deep Learning and the large availability of data and GPUs, great strides have been made in predicting the memorability of an image. In this paper, we propose a novel deep learning architecture called ResMem-Net that is a hybrid of LSTM and CNN that uses information from the hidden layers of the CNN to compute the memorability score of an image. The intermediate layers are important for predicting the output because they contain information about the intrinsic properties of the image. The proposed architecture automatically learns visual emotions and saliency, shown by the heatmaps generated using the GradRAM technique. We have also used the heatmaps and results to analyze and answer one of the most important questions in image memorability: ‘‘What makes an image memorable?“. The model is trained and evaluated using the publicly available Large-scale Image Memorability dataset (LaMem) from MIT. The results show that the model achieves a rank correlation of 0.679 and a mean squared error of 0.011, which is better than the current state-of-the-art models and is close to human consistency (p = 0.68). The proposed architecture also has a significantly low number of parameters compared to the state-of-the-art architecture, making it memory efficient and suitable for production.
Impact and Reach
Statistics
Additional statistics for this dataset are available via IRStats2.