Skip to content

Releases: google-deepmind/optax

Optax 0.1.4.

21 Nov 13:44
Compare
Choose a tag to compare

What's Changed

New features

  • Expose adamax and adamaxw by @nasyxx in #433
  • Support feeding in extra dictionary data in optax/experimental in #373
  • Add NovoGrad optimizer by @DT6A in #385
  • added optimistic gradient descent in #387
  • Add types to utils.py by @atgctg in #367
  • Add hyperparam_dtype override to hyperparam injection in #414
  • added typing to linear_algebra.py by @shivance in #413
  • Add amsgrad optimizer by @merajhashemi in #382
  • Add Hinge Loss by @heytanay in #409

Bug fixes

  • [optax] Increase chex version requirement to fix issue #456. in #457
  • Start inject_hyperparams step count at 0. in #416
  • Add noise before multiplying by the learning rate (noisy_sgd ) by @atgctg in #369

Miscellaneous

  • Remove flags from lookahead_mnist example. in #395
  • Bring alias docstrings more in line with style guide. in #406
  • Bugfix in alias_test: Adamax named but not tested. in #419
  • Test that all optimizers can be wrapped in inject_hyperparams. in #420
  • Add an example for gradient accumulation. in #425
  • [docs] Start adding numerical definitions to key parts of optax. in #430
  • [optax] Add basic mnist example based on the lookahead_mnist example. in #436
  • Install dependencies for dp-accounting manually. in #375
  • Update documentation for AdamW in #376
  • [Docs] softmax_cross_entropy_with_integer_label by @lkhphuc in #351
  • Update Returns section in gradient transformations' docstrings. in #388
  • Update logo and theme for the documentation in #391
  • Set the test version of flax used in the equivalence test. in #398
  • Add holounic to contributors list. by in #400
  • Bring transform docstrings more in line with style guide. by in #405
  • Update citation. in #407
  • Refine the doc of sigmoid_binary_cross_entropy to not assume the meaning of last dimension. in #418
  • fix integer_pow recompilation in _bias_correction by @epignatelli in #329
  • Use auto instead of /proc/cpuinfo. in #454

New Contributors

Full Changelog: v0.1.3...v0.1.4

Optax 0.1.3.

13 Jul 11:28
Compare
Choose a tag to compare

What's Changed

  • Do not use None in optax.masked. in #338
  • Implement Adamax and AdamaxW optimizers. in #342
  • Add Kullback-Leibler Divergence Loss by @holounic in #309
  • Add optax.softmax_cross_entropy_with_integer_labels. in #343
  • Publicize private methods in transform in #348
  • Update .pylintrc in #354
  • Support mixture of dtypes for parameters when clipping. in #355
  • Update "jax.tree_util" functions in #370

New Contributors

Full Changelog: v0.1.2...v0.1.3

Optax 0.1.2

13 Apr 10:21
Compare
Choose a tag to compare

What's Changed

  • Improve the documentation and discoverability of set_to_zero. by @copybara-service in #299
  • Add pages to documentation for contributors. by @copybara-service in #301
  • Replace _build_simple_adam with _build_stateful_sgd. by @n2cholas in #302
  • Make masked preserve param structure by @n2cholas in #300
  • Allow to set custom schedule for the second momentum in Adafactor. One useful use case is setting the upper bound for the second momentum, as otherwise it converges to 1.0 and effectively freezes the second momentum. It was used in https://arxiv.org/abs/2106.04560. by @copybara-service in #303
  • exporting CTC loss for public use by @copybara-service in #305
  • Specify that mask must be static in masked wrapper docstring. by @copybara-service in #306
  • Clarify the docstring of cosine_decay_schedule. by @copybara-service in #310
  • Typo in LARS docstring. by @Olamon in #311
  • Fix adam's mu_dtype casting. by @copybara-service in #313
  • Fix docs rendering of loss functions. by @grahamgower in #318
  • Clarifies optax.adamw(weight_decay) parameter. by @copybara-service in #322
  • Enhance ctc_loss to return forward probs. by @yotarok in #321
  • Replace jax.tree_multimap, which is now deprecated, with jax.tree_map. by @copybara-service in #330
  • Rename argument names for CTC loss functions. by @yotarok in #331

New Contributors

Full Changelog: v0.1.1...v0.1.2

Optax 0.1.1

04 Feb 08:24
Compare
Choose a tag to compare

What's Changed

  • Tweak the meta-learning example from the docs by @copybara-service in #233
  • Fix small bugs in metalearning example. by @copybara-service in #236
  • Do not reuse mini-batches between epochs in DPSGD example. by @copybara-service in #230
  • Make the version of typing_extensions less constrained. by @copybara-service in #238
  • [JAX] move example libraries from jax.experimental to jax.example_libraries by @copybara-service in #200
  • Export ScaleByBeliefState by @NeilGirdhar in #239
  • MultiStep optimizer: align parameter naming and type annotations of update function with signature of GradientTransform.update. by @copybara-service in #243
  • Fix imports of datasets in examples folder. by @copybara-service in #242
  • Enable example tests on github. Fix the bugs that were uncovered. by @copybara-service in #244
  • Formatting. by @copybara-service in #249
  • Add test for multi steps wrapper, verifying that the aggregated gradient is the mean of the input gradients. by @copybara-service in #255
  • MultiStep optimizer wrapper: replace naive streaming average gradient implementation with numerically stabler one. by @copybara-service in #254
  • Added ord, axis, and keepdims args to safe_norm by @copybara-service in #252
  • Add badges and include RTD build into CI tests. by @copybara-service in #256
  • Write a clearer doc-string for GradientTransformation by @copybara-service in #257
  • Refactor clipping.py by @copybara-service in #260
  • Implement split real norm by @wdphy16 in #241
  • Monkey-patch sphinx to output correct type annotations for the most common cases (e.g. params, opt state) in the documentation. by @copybara-service in #266
  • Improve docs by @copybara-service in #268
  • Implement stateless wrapper. by @n2cholas in #246
  • Replace _ with params to ensure you can always call init with named args. by @copybara-service in #270
  • Improve docs. by @copybara-service in #269
  • Add missing ` in two places. by @copybara-service in #273
  • Add option to cache examples datasets after pre-processing. by @copybara-service in #272
  • Fix an error in README.md rendering. by @copybara-service in #275
  • Remove the old venv directory before testing the package. by @copybara-service in #289
  • Fix Yogi optimizer by @wdphy16 in #288
  • Bump ipython from 7.16.1 to 7.16.3 in /requirements by @dependabot in #286
  • Clarifies optax.adamw(mask) parameter. by @copybara-service in #284
  • Fix the link to the complex-valued optim proposal in RTD. by @copybara-service in #295
  • Implement complex norm in optimizers by @wdphy16 in #279
  • Change add_noise to match the target variance by scaling by its sqrt. by @Rupt in #294
  • Minor tweaks to the optax documentation. by @copybara-service in #297
  • Bump version to 0.1.1 from 0.1.0 by @copybara-service in #298

New Contributors

Full Changelog: v0.1.0...v0.1.1

Optax 0.1.0

18 Nov 19:51
Compare
Choose a tag to compare

Support for Python 3.6 has been dropped following the JAX deprecation policy. Please upgrade to a supported Python version.

Full Changelog: v0.0.91...v0.1.0

Optax 0.0.91

18 Nov 18:34
Compare
Choose a tag to compare

It is the latest version compatible with Python 3.6. See #222 for more details.

Main changes since 0.0.9

  • Use an explicit type for base classes of optimizer states. #211
  • Replace optax.OptState with NamedTuple, when it is used as a base class. #218
  • Set optax.OptState to chex.ArrayTree. #219
  • Add quick start notebook to Optax. #143
  • Add Optax regression flax_example.py. by @8bitmp3 in #138
  • Add scale_gradient. #198
  • Add a set_to_zero transformation. #202
  • Add utilities for numerical stability in the public API. #227
  • Change int64 to safely incremented int32 in wrappers. by @rdaems in #187
  • Add an example of meta-optimization to the docs. #231

Full Changelog: v0.0.9...v0.0.91

Optax 0.0.9

02 Jul 17:56
Compare
Choose a tag to compare

Added:

  • multi_transform
  • LARS optimiser
  • AdaFactor optimiser

Fixed:

  • masked wrapper for empty params nest
  • label type in loss.py
  • checks in loss.py
  • MultiSteps arguments handling
  • Chex asserts in constrain_test.py

Optax 0.0.8

10 Jun 12:50
Compare
Choose a tag to compare
  • Added clipping_by_block_rms
  • Added sgdr_schedule
  • Fixed inconsistency in ema's outputs
  • Added linear_schedule

Optax 0.0.6

26 Mar 12:31
Compare
Choose a tag to compare

Changelog

Full Changelog

Closed issues:

  • Documentation #3

Merged pull requests:

* This Changelog was automatically generated by github_changelog_generator

Optax 0.0.5

23 Mar 18:42
Compare
Choose a tag to compare

Changelog

Note: this is a first GitHub release of Optax. It includes all changes since the repo was created.

Full Changelog

Implemented enhancements:

  • Implement lookahead optimiser #17
  • Implement support for Yogi optimiser #9
  • Implement rectified Adam #8
  • Implement gradient centralisation #7
  • Implement scaling by AdaBelief #6

Closed issues:

  • Multiple optimizers using optax #59
  • Change masked wrapper to use mask_fn instead of mask #57
  • Prevent creating unnecessary momentum variables #52
  • Implement Differentially Private Stochastic Gradient Descent #50
  • RMSProp does not match original Tensorflow impl #49
  • JITted Adam results in NaN when setting decay to integer 0 #46
  • Option to not decay bias with additive_weight_decay #25
  • Support specifying end_value for exponential_decay #21
  • Schedules for Non-Learning Rate Hyper-parameters #20
  • Implement OneCycle Learning Rate Schedule #19
  • adam does not learn? #18
  • Which JAX-based libraries is optax compatible with? #14
  • Manually setting the learning_rate? #4

Merged pull requests:

* This Changelog was automatically generated by github_changelog_generator