Skip to content

Commit

Permalink
update: disable coverage
Browse files Browse the repository at this point in the history
  • Loading branch information
kozistr committed Jun 27, 2024
1 parent beb84d4 commit bede547
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions pytorch_optimizer/optimizer/adam_mini.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
from pytorch_optimizer.base.types import BETAS, CLOSURE, DEFAULTS, LOSS


class AdamMini(Optimizer, BaseOptimizer):
class AdamMini(Optimizer, BaseOptimizer): # pragma: no cover
r"""Use Fewer Learning Rates To Gain More.
:param model: nn.Module. model instance.
Expand Down Expand Up @@ -218,7 +218,7 @@ def step_lefts(
bias_correction1: float,
bias_correction2_sq: float,
eps: float,
) -> None: # pragma: no cover
) -> None:
if len(state) == 0:
dim = torch.tensor(p.numel(), device=p.device, dtype=torch.float32)

Expand Down

0 comments on commit bede547

Please sign in to comment.