optimizer_base_36e.yml 310 B

1234567891011121314151617181920
  1. epoch: 36
  2. LearningRate:
  3. base_lr: 0.0001
  4. schedulers:
  5. - !CosineDecay
  6. max_epochs: 36
  7. min_lr_ratio: 0.1 # 0.1
  8. - !LinearWarmup
  9. start_factor: 0.001
  10. epochs: 1
  11. OptimizerBuilder:
  12. clip_grad_by_norm: 0.1
  13. regularizer: false
  14. optimizer:
  15. type: AdamW
  16. weight_decay: 0.0001