Replies: 2 comments 1 reply
-
@mmaaz60 not sure if it's 100% exact, but would be close to https://gist.github.com/rwightman/943c0fe59293b44024bbd2d5d23e6303#file-maxvit_tiny_256-yaml or the nano there... although this was for a 256x256, don't think I ever finished it, but would have been based on the 224 |
Beta Was this translation helpful? Give feedback.
1 reply
-
Hi @mmaaz60 Here is my training receipt to train torchrun --nproc_per_node=4 --master_port=12345 train.py /path/to/imagenet --model maxvit_tiny_tf_224 --aa rand-m15-mstd0.5-inc1 --mixup .8 --cutmix 1.0 --remode pixel --reprob 0.25 --drop-path .2 --opt adamw --weight-decay .05 --sched cosine --epochs 300 --lr 3e-3 --warmup-lr 1e-6 --warmup-epoch 30 --min-lr 1e-5 -b 64 --grad-accum-steps 16 --smoothing 0.1 --clip-grad 1.0 -j 8 --amp --pin-mem --channels-last I also upload the training logs and checkpoints at this repo. Thank you. Hankyul |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
The MaxViT-tiny model (maxvit_tiny_rw_224) provided in Timm achieves 83.51 top-1 IN1K accuracy when evaluated. However, I am interested in training this model in Timm and reproduce the same numbers. Can you share the exact training recipe (ideally config/training command) to reproduce these numbers in timm?
Thanks
Beta Was this translation helpful? Give feedback.
All reactions