TanAlexanderlz

TanAlexanderlz/ALL_RGBCROP_ori16F-8B16F-GACWD1

More information needed...

Model Documentation

This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. -->

ALL_RGBCROP_ori16F-8B16F-GACWD



This model is a fine-tuned version of MCG-NJU/videomae-base-finetuned-kinetics on an unknown dataset. It achieves the following results on the evaluation set:
  • Loss: 0.3803
  • Accuracy: 0.8144


  • Model description



    More information needed

    Intended uses & limitations



    More information needed

    Training and evaluation data



    More information needed

    Training procedure



    Training hyperparameters



    The following hyperparameters were used during training:
  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 16
  • optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • training_steps: 1440


  • Training results



    | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:------:|:----:|:---------------:|:--------:| | 0.6146 | 0.0333 | 48 | 0.6323 | 0.6280 | | 0.3307 | 1.0333 | 96 | 0.4748 | 0.7805 | | 0.2425 | 2.0333 | 144 | 0.6149 | 0.7805 | | 0.1629 | 3.0333 | 192 | 0.7126 | 0.7683 |

    Framework versions



  • Transformers 4.51.3
  • Pytorch 2.6.0+cu124
  • Datasets 2.14.4
  • Tokenizers 0.21.1
  • Files & Weights

    FilenameSizeAction
    model.safetensors 0.32 GB
    training_args.bin 0.00 GB