-
Notifications
You must be signed in to change notification settings - Fork 467
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
YOLO-WORLD-S在coco上finetune无法复现,且validation map呈现下降趋势 #160
Comments
Hi @shupinghu, we have also met the same problem: fine-tuning YOLO-World on COCO without |
Does the validation map in your experiment gradually decrease during fine-tuning as well? In my experiment, not only was map unable to reproduce the values in the paper, but the bigger problem was that map was getting worse and worse during fine-tuning. |
@shupinghu, |
OK, I will try this config file and feedback the experiment result. Does use "mask-refine" mean that we use the segmentation annotation to reproduce the bbox annoration? |
|
|
|
@wondervictor |
@taofuyu I'll add it in TODO and fix it soon. |
Hi @shupinghu and @taofuyu, I've uploaded the fine-tuned weights and logs for models with |
using "mask-refine" is OK. |
Updates: the performance will be much worse without the |
[Failed Update] : using optim_wrapper = dict(optimizer=dict(
_delete_=True,
type='SGD',
lr=1e-3,
momentum=0.937,
nesterov=True,
weight_decay=0.0005,
batch_size_per_gpu=train_batch_size_per_gpu)) |
Hi all (@taofuyu, @shupinghu): happy to update a milestone,
max_epochs = 40 # Maximum training epochs
close_mosaic_epochs = 30
optim_wrapper = dict(
optimizer=dict(_delete_=True,
type='SGD',
lr=1e-3,
momentum=0.937,
nesterov=True,
weight_decay=weight_decay,
batch_size_per_gpu=train_batch_size_per_gpu),
paramwise_cfg=dict(custom_keys={'logit_scale': dict(weight_decay=0.0)}),
constructor='YOLOWv5OptimizerConstructor') Under this setting, YOLO-World-Large without This is a milestone but not the terminus and we are still working on it for a better fine-tuning setting! Those updates will be pushed in a day. |
Hi all (@taofuyu, @shupinghu), we have preliminarily explored the errors about pre-training without You can find more details in configs/finetune_coco, especially for the version without |
Hi @wondervictor, I met the same issues for fine-tuning on my own dataset. The mAP50 will decrease after 15 epochs. Do you have any idea about that? I have tried the two fine-tuning config files yolo_world_l_dual_vlpan_2e-4_80e_8gpus_finetune_coco.py and yolo_world_v2_l_vlpan_bn_2e-4_80e_8gpus_finetune_coco_womixup.py (I think you delete this file in the current version). Both are decrest after 15 epochs. However, performance will increase in the last 10 epochs. I just modified the img_scale to 1280, 960 and max_epochs to 100. Other parameters are the same as your configs. Unfortunately, my dataset does not include the mask, so I can not use mask-refine. |
Hi @JiayuanWang-JW, could you try out the latest config for your custom data? I've preliminarily fixed the above issues. The new config does not require |
Thanks for your rapid response. I have finished the experiment on my own dataset. It is much better than before config. The current result as shown If I continue fine-tuning more epochs, I need to change the close_mosaic_epochs and which parameters (such as base_lr, weight_decay, etc)? I want to try 100 epochs. The best mAP50 is 0.607. I think this is not enough, to be honest, some classical detection methods are much better than this value. And the number of parameters is much less than YOLO-World-L. Do you have any idea how to continue to improve the performance? |
Hi @JiayuanWang-JW, is there any update? |
Hi @wondervictor. Thanks for your reply. No. I didn't get the better YOLO-World results on my dataset. Actually, I already used (1280, 960) size and different learning rate strategies(such as Cosine Annealing and different T_max) to fine-tune the detection tasks. I tried 100 epochs and different mosaic epochs. However, the best performance only achieved 60.7% for AP50. After that, the performance will decrease. Just select two examples to explain. YOLOv8 achieved 75.1% on my dataset. They are a large gap. Anyway, I will continue to explore it and update them here if I find any useful information. |
【复现步骤】
你好,我下载了yolo_world_s_clip_base_dual_vlpan_2e-3adamw_32xb16_100e_o365_goldg_train_pretrained-18bea4d2.pth,加载它进行预训练,config文件是从configs/finetune_coco/yolo_world_l_dual_vlpan_2e-4_80e_8gpus_finetune_coco.py拷贝后修改的,即将_base_里面的yolov8_l_syncbn_fast_8xb16-500e_coco.py直接修改为yolov8_s_syncbn_fast_8xb16-500e_coco.py,然后在yolov8_s_syncbn_fast_8xb16-500e_coco.py里增加了“mixup_prob = 0.1”以解决mmengine报错的问题。
此外,我只用了单卡V100进行训练,担心单卡比8卡batch size差距太大,我将单卡的batch size从16修改成了32。
【实验现象】
其他没有做修改,开始训练后发现validation map从第15个epoch开始,就呈现下降趋势(从epoch 10的0.414、0.576下降到epoch 15的0.410、0.574,当前训练到epoch55,已经下降到了0.377、0.541)。
【问题咨询及分析】
想请问是否是我哪里复现有问题?(check了一下貌似直接改成s之后mixup好像没有开,有可能是这个原因,但是我觉得不至于会让validation map呈下降趋势)
The text was updated successfully, but these errors were encountered: