An efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
-
Updated
Mar 10, 2024 - Python
An efficient PyTorch implementation of the winning entry of the 2017 VQA Challenge.
A PyTorch reimplementation of bottom-up-attention models
Image captioning with Transformer
Visual Question Answering using Transformer and Bottom-Up attention. Implemented in Pytorch
Extract features and bounding boxes using the original Bottom-up Attention Faster-RCNN in a few lines of Python code
The main goal of is to show how precise the Faster R-CNN with ResNet-101 could find objects and there attributes in Conceptual 12m dataset.
Add a description, image, and links to the bottom-up-attention topic page so that developers can more easily learn about it.
To associate your repository with the bottom-up-attention topic, visit your repo's landing page and select "manage topics."