summaryrefslogtreecommitdiff
path: root/torch/optim
diff options
context:
space:
mode:
authorSoumith Chintala <soumith@gmail.com>2018-10-24 09:01:12 -0700
committerFacebook Github Bot <facebook-github-bot@users.noreply.github.com>2018-10-24 09:03:38 -0700
commitcf235e0894eb78b741afb98f15da62673ab41dfe (patch)
treefc90b1f241178aa601ef41ae7b2ce92780d50cad /torch/optim
parentd72de9fb1e6e8547d772a901ddbe1ec8a34b4c10 (diff)
downloadpytorch-cf235e0894eb78b741afb98f15da62673ab41dfe.tar.gz
pytorch-cf235e0894eb78b741afb98f15da62673ab41dfe.tar.bz2
pytorch-cf235e0894eb78b741afb98f15da62673ab41dfe.zip
fix lint after new flake8 release added new style constraints (#13047)
Summary: fix lint after new flake8 release added new style constraints Pull Request resolved: https://github.com/pytorch/pytorch/pull/13047 Differential Revision: D10527804 Pulled By: soumith fbshipit-source-id: 6f4d02662570b6339f69117b61037c8394b0bbd8
Diffstat (limited to 'torch/optim')
-rw-r--r--torch/optim/adam.py2
-rw-r--r--torch/optim/sparse_adam.py2
2 files changed, 2 insertions, 2 deletions
diff --git a/torch/optim/adam.py b/torch/optim/adam.py
index a26de99ec0..edcfcc26be 100644
--- a/torch/optim/adam.py
+++ b/torch/optim/adam.py
@@ -4,7 +4,7 @@ from .optimizer import Optimizer
class Adam(Optimizer):
- """Implements Adam algorithm.
+ r"""Implements Adam algorithm.
It has been proposed in `Adam: A Method for Stochastic Optimization`_.
diff --git a/torch/optim/sparse_adam.py b/torch/optim/sparse_adam.py
index afb23bfa46..252645794c 100644
--- a/torch/optim/sparse_adam.py
+++ b/torch/optim/sparse_adam.py
@@ -4,7 +4,7 @@ from .optimizer import Optimizer
class SparseAdam(Optimizer):
- """Implements lazy version of Adam algorithm suitable for sparse tensors.
+ r"""Implements lazy version of Adam algorithm suitable for sparse tensors.
In this variant, only moments that show up in the gradient get updated, and
only those portions of the gradient get applied to the parameters.