Skip to content

Latest commit

 

History

History
38 lines (29 loc) · 1.42 KB

CHANGELOG.md

File metadata and controls

38 lines (29 loc) · 1.42 KB

Changelog

1.0.8

  • Removed redundant code from reduce_precision.
  • Added types argument to PseudoParameter.parametrize_module for better selection for Parameterising the Layers.

1.0.7

  • Fixed GeLU backward function equation.

1.0.6

  • Model is subclass of BackwardModule for additional functionality.
  • Using inspect.isclass to check if backward_class is a class in Linear.set_backward_function.
  • Repr using self.__class__.__name__ in all classes.

1.0.5 (Patches for Pytorch 2.0.1)

  • Removed unnecessary PseudoParameter.grad property.
  • Patch for Pytorch 2.0.1, add filtering inputs in BackwardGraph._calculate_gradients.

1.0.4

  • Combined PseudoParameter and PseudoParameterModule for better visibility.
    • BugFix: fixed save and load of state_dict of PseudoParameter and transformation module.
  • Removed redundant class analogvnn.parameter.Parameter.

1.0.3

  • Added support for no loss function in Model class.
    • If no loss function is provided, the Model object will use outputs for gradient computation.
  • Added support for multiple loss outputs from loss function.

1.0.2

  • Bugfix: removed graph from Layer class.
    • graph was causing issues with nested Model objects.
    • Now _use_autograd_graph is directly set while compiling the Model object.

1.0.1 (Patches for Pytorch 2.0.0)

  • added grad.setter to PseudoParameterModule class.

1.0.0

  • Public release.