Included yaml files are timm train script configs for training MobileNetV4 models in timm (see on HF Hub: https://huggingface.co/collections/timm/mobilenetv4-pretrained-weights-6669c22cda4db4244def9637)
Note the # of GPUs, this needs to be taken into consideration for global batch size equivalence, and LR scaling.
Also note, some models have lr
set to a non null value, this LR is used directly if set. Otherwise, it falls back to lr_base
and the used rate is calculated based on lr_base_sized
and a sqrt scaling according to the global batch size.
Thank you for sharing.