Skip to content

Releases: kozistr/pytorch_optimizer

pytorch-optimizer v2.11.1

19 Jul 12:00
aaaf303
Compare
Choose a tag to compare

pytorch-optimizer v2.11.0

27 Jun 13:48
42df19c
Compare
Choose a tag to compare

pytorch-optimizer v2.10.1

13 Jun 06:37
baa65c4
Compare
Choose a tag to compare

Change Log

Feature

Fix

  • perturb isn't multiplied by -step_size in SWATS optimizer. (#179)
  • chebyshev step has size of T while the permutation is 2^T. (#168, #181)

Diff

2.10.0...2.10.1

pytorch-optimizer v2.10.0

07 Jun 14:18
95eee86
Compare
Choose a tag to compare

Change Log

Feature

Diff

2.9.1...2.10.0

Contributions

thanks to @i404788

pytorch-optimizer v2.9.1

19 May 12:09
9427d3c
Compare
Choose a tag to compare

Change Log

Fix

  • fix weight decay in Ranger21 (#170)

Diff

2.9.0...2.9.1

pytorch-optimizer v2.9.0

06 May 08:07
4dbfc23
Compare
Choose a tag to compare

Change Log

Feature

Docs

  • Fix readthedocs build issue, #156
  • Move citations into table, #156

Refactor

  • Refactor validation logic, #149, #150
  • Rename amsbound, amsgrad terms into ams_bound, #149
  • Return gradient instead of the parameter, AGC. #149
  • Refactor duplicates (e.g. rectified step size, AMSBound, AdamD, AdaNorm, weight decay) into re-usable functions, #150
  • Move pytorch_optimizer.experimental under pytorch_optimizer.*.experimental

Diff

2.8.0...2.9.0

pytorch-optimizer v2.8.0

29 Apr 08:51
cdfe807
Compare
Choose a tag to compare

pytorch-optimizer v2.7.0

26 Apr 06:31
7ded073
Compare
Choose a tag to compare

Change Log

Features

Refactor

  • Rename adamd_debias_term to adam_debias, #133
  • Merge the rectified version with the original, #133
    • diffRGrad + diffGrad -> diffGrad
    • RaLamb + Lamb -> Lamb
    • now you can simply use with rectify=True

Fix

  • Fix previous_grad deepcopy issue in Adan optimizer. #134

pytorch-optimizer v2.6.1

22 Apr 12:14
be0351d
Compare
Choose a tag to compare

Change Log

Fix

  • variables are not located on the same device with the gradients, #132 (related to #131) (thanks to @Bing-su)
  • fix approximate_sq_grad() in Adafactor optimizer, #132

pytorch-optimizer v2.6.0

22 Apr 07:56
19dcf2b
Compare
Choose a tag to compare

Change Log

Feature

  • Implement SM3 optimizer, #130
  • Tweak Scalable Shampoo optimizer, #128, #129
    • implement a new preconditioner type, OUTPUT.
    • optimize speed/memory usage of coupled Newton iteration and power iteration methods.
    • use in-place operation (SQRT-N Grafting).
    • clean-up shampoo_utils more readable.
    • support skip_preconditioning_rank_lt parameter to skip preconditioning in case of the low-rank gradient.
    • set default value for preconditioning_compute_steps to 1000.
    • set default value for start_preconditioning_step to 25.