Skip to content

Releases: NTT123/opax

v0.2.11

30 Jul 10:24
Compare
Choose a tag to compare

v0.2.10

19 Jun 06:46
Compare
Choose a tag to compare
  • ScaleBySchedule: correct the learning rate at initialization.

Full Changelog: v0.2.8...v0.2.10

v0.2.8

25 Nov 09:35
Compare
Choose a tag to compare

Introduce strict and non-strict modes:

  • Strict mode requires pytrees to be matched.
  • Non-strict mode only requires tree leaves to be matched.

v0.2.7

21 Nov 13:50
Compare
Choose a tag to compare
  • Make testcases compatible with pax3>=0.5.0.

v0.2.6

18 Nov 17:29
Compare
Choose a tag to compare
  • Simplify the implementation by considering gradient transformations as StateModule.

v0.2.5

13 Nov 18:30
Compare
Choose a tag to compare

Introduce the class method init. It makes things easier to understand because we explicitly say "initialize the optimizer".

Implicit:

optimizer = opax.adam(1e-3)(net.parameters())

Explicit:

optimizer = opax.adam(1e-3).init(net.parameters())

v0.2.4

19 Oct 15:33
Compare
Choose a tag to compare
  • Import add_decayed_weights.
  • Add __all__ to __init__.py file.

v0.2.3

13 Oct 14:19
c89d530
Compare
Choose a tag to compare
  • Use the newest PAX v0.4.0.
  • Remove step method.
  • Add transform_gradients, apply_updates and apply_gradients utility (pure) functions.

v0.1.7

21 Sep 03:50
Compare
Choose a tag to compare
  • Check if param's tree structure matchs gradient's tree structure for better error messages.
  • Support all_finite predicate when updating parameters. This is useful when models are trained with mixed precision.
  • Support flatten mode to speed up tree_flatten and tree_unflatten.

v0.1.0

08 Sep 06:55
92e9214
Compare
Choose a tag to compare

First official relase:

  • support adam, adamw, rmsprop and sgd optimizers.
  • support lr schedule.