User Constrained Thumbnail Generation using Adaptive Convolutions [paper]
The paper proposes a deep neural framework to generate thumbnails of any size and aspect ratio, even for unseen values during training, with high accuracy and precision. It uses Global Context Aggregation (GCA) and a modified Region Proposal Network (RPN) with adaptive convolutions to generate thumbnails in real time. GCA is used to selectively attend and aggregate the global context information from the entire image while the RPN is used to predict candidate bounding boxes for the thumbnail image. Adaptive convolution eliminates the problem of generating thumbnails of various aspect ratios by using filter weights dynamically generated from the aspect ratio information.
The original image is shown on the left with the generated thumbnails on the right. The query aspect ratio is given in blue and the aspect ratio of the generated thumbnail is given in red.
The code is written in Python 3 with Tensorflow 1.6. The pretrained weights for the custom VGG-19 model can be found here.
The dataset used can be found here.
If you find this work useful in your research, please consider citing:
@article{kishore2018user,
title={User Constrained Thumbnail Generation using Adaptive Convolutions},
author={Kishore, Perla Sai Raj and Bhunia, Ayan Kumar and Ghose, Shuvozit and Roy, Partha Pratim},
journal={arXiv preprint arXiv:1810.13054},
year={2018}
}
This paper is accepted to IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), 2019.