This repo holds the pytorch-version codes of paper: "BSN: Boundary Sensitive Network for Temporal Action Proposal Generation", which is accepted in ECCV 2018. You can also find tensorflow-version implementation in [BSN.tensorflow].
- 2018.12.12: Release Pytorch-version BSN
- 2018.09.26: Previously, we adopted classification results from result files of "Uts at activitynet 2016" for action detection experiments. Recently we found that the classification accuracy of these results are unexpected high. Thus we replace it with classification results of "cuhk & ethz & siat submission to activitynet challenge 2017" and updated all related experiments accordingly. You can find updated papers in my homepage and in arXiv.
- 2018.07.09: Codes and feature of BSN
- 2018.07.02: Repository for BSN
Temporal action proposal generation is an important yet challenging problem, since temporal proposals with rich action content are indispensable for analysing real-world videos with long duration and high proportion irrelevant content. This problem requires methods not only generating proposals with precise temporal boundaries, but also retrieving proposals to cover truth action instances with high recall and high overlap using relatively fewer proposals. To address these difficulties, we introduce an effective proposal generation method, named Boundary-Sensitive Network (BSN), which adopts “local to global” fashion. Locally, BSN first locates temporal boundaries with high probabilities, then directly combines these boundaries as proposals. Globally, with Boundary-Sensitive Proposal feature, BSN retrieves proposals by evaluating the confidence of whether a proposal contains an action within its region. We conduct experiments on two challenging datasets: ActivityNet-1.3 and THUMOS14, where BSN outperforms other state-of-the-art temporal action proposal generation methods with high recall and high temporal precision. Finally, further experiments demonstrate that by combining existing action classifiers, our method significantly improves the state-of-the-art temporal action detection performance.
These code is implemented in Pytorch 0.4.1 + Python2 + tensorboardX. Thus please install Pytorch first.
Clone this repo with git, please use:
git clone https://github.com/wzmsltw/BSN-boundary-sensitive-network.pytorch.git
We support experiments with publicly available dataset ActivityNet 1.3 for temporal action proposal generation now. To download this dataset, please use official ActivityNet downloader to download videos from the YouTube.
To extract visual feature, we adopt TSN model pretrained on the training set of ActivityNet, which is the challenge solution of CUHKÐ&SIAT team in ActivityNet challenge 2016. Please refer this repo TSN-yjxiong to extract frames and optical flow and refer this repo anet2016-cuhk to find pretrained TSN model.
For convenience of training and testing, we rescale the feature length of all videos to same length 100, and we provide the rescaled feature at here Google Cloud or Baidu Yun. If you download features using BaiduYun, please use cat zip_csv_mean_100.z* > csv_mean_100.zip
before unzip. After download and unzip, please put csv_mean_100
directory to ./data/activitynet_feature_cuhk/
.
All configurations of BSN are saved in opts.py, where you can modify training and model parameter.
python main.py --module TEM --mode train
We also provide trained TEM model in ./checkpoint/
python main.py --module TEM --mode inference
python main.py --module PGM
python main.py --module PEM --mode train
We also provide trained PEM model in ./checkpoint
.
python main.py --module PEM --mode inference --pem_batch_size 1
python main.py --module Post_processing
python main.py --module Evaluation
You can find evaluation figure in ./output
You can also simply run all above commands using:
sh bsn.sh
Please cite the following paper if you feel SSN useful to your research
@inproceedings{BSN2018arXiv,
author = {Tianwei Lin and
Xu Zhao and
Haisheng Su and
Chongjing Wang and
Ming Yang},
title = {BSN: Boundary Sensitive Network for Temporal Action Proposal Generation},
booktitle = {European Conference on Computer Vision},
year = {2018},
}
For any question, please file an issue or contact
Tianwei Lin: wzmsltw@sjtu.edu.cn