optimizer_1x.yml 254 B

12345678910111213141516
  1. epoch: 12
  2. LearningRate:
  3. base_lr: 0.0001
  4. schedulers:
  5. - !PiecewiseDecay
  6. gamma: 0.1
  7. milestones: [11]
  8. use_warmup: false
  9. OptimizerBuilder:
  10. clip_grad_by_norm: 0.1
  11. regularizer: false
  12. optimizer:
  13. type: AdamW
  14. weight_decay: 0.0001