Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Test] Add more test cases #45

Merged
merged 16 commits into from
Jan 29, 2022
Prev Previous commit
Next Next commit
update: RAdam
  • Loading branch information
kozistr committed Jan 29, 2022
commit 6fb3a4f25ec62a1c50aa61808b95006a8372bf95
6 changes: 3 additions & 3 deletions pytorch_optimizer/radam.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,13 +35,13 @@ def __init__(
adamd_debias_term: bool = False,
eps: float = 1e-8,
):
"""
"""RAdam
:param params: PARAMETERS. iterable of parameters to optimize or dicts defining parameter groups
:param lr: float. learning rate.
:param lr: float. learning rate
:param betas: BETAS. coefficients used for computing running averages of gradient and the squared hessian trace
:param weight_decay: float. weight decay (L2 penalty)
:param n_sma_threshold: int. (recommended is 5)
:param degenerated_to_sgd: float.
:param degenerated_to_sgd: float. degenerated to SGD
:param adamd_debias_term: bool. Only correct the denominator to avoid inflating step sizes early in training
:param eps: float. term added to the denominator to improve numerical stability
"""
Expand Down