Muzhi Zhu1*, Chengxiang Fan1*, Hao Chen1†, Yang Liu1, Weian Mao2 1, Xiaogang Xu3 1, Chunhua Shen1†
1 Zhejiang University, 2 The University of Adelaide, 3 The Chinese University of Hong Kong
* Equal contribution † Corresponding author
Recently, large-scale language-image generative models have gained widespread attention and many works have utilized generated data from these models to further enhance the performance of perception tasks. However, not all generated data can positively impact downstream models, and these methods do not thoroughly explore how to better select and utilize generated data. On the other hand, there is still a lack of research oriented towards active learning on generated data. In this paper, we explore how to perform active learning specifically for generated data in the long-tailed instance segmentation task. Subsequently, we propose BSGAL, a new algorithm that online estimates the contribution of the generated data based on gradient cache. BSGAL can handle unlimited generated data and complex downstream segmentation tasks effectively. Experiments show that BSGAL outperforms the baseline approach and effectually improves the performance of long-tailed segmentation.
- 2024.7 We have released the code of BSGAL!
- 2024.5 BSGAL has been accepted to ICML 2024!
- Release base codes
- Release weights
Prepare the environment and dataset following X-Paste.
bash launch.sh --config configs/BSGAL/BSGAL_R50.yaml
First download the weights from Here
bash launch.sh --config configs/BSGAL/BSGAL_R50.yaml --eval-only MODEL.WEIGHTS path/to/weights
We thank the following repos for their great works:
If you find this project useful in your research, please cite:
@inproceedings{zhu2024generative,
title={Generative Active Learning for Long-tailed Instance Segmentation},
author={Muzhi Zhu and Chengxiang Fan and Hao Chen and Yang Liu and Weian Mao and Xiaogang Xu and Chunhua Shen},
booktitle={International Conference on Machine Learning},
year={2024},
}