diff options
author | Edward Yang <ezyang@fb.com> | 2019-03-14 08:52:55 -0700 |
---|---|---|
committer | Facebook Github Bot <facebook-github-bot@users.noreply.github.com> | 2019-03-14 08:59:24 -0700 |
commit | 3fe7bdb2ff4b1c49db69fad61d3f6f03702cf0d6 (patch) | |
tree | 986e5e4bd967c465160b0ff13a05691e102d4c05 /test/test_nn.py | |
parent | a41b6d7d1f96a432ef6d2b75a25a2db5ec7ffbbb (diff) | |
download | pytorch-3fe7bdb2ff4b1c49db69fad61d3f6f03702cf0d6.tar.gz pytorch-3fe7bdb2ff4b1c49db69fad61d3f6f03702cf0d6.tar.bz2 pytorch-3fe7bdb2ff4b1c49db69fad61d3f6f03702cf0d6.zip |
Fix lint in test_nn.py (#18006)
Summary:
Pull Request resolved: https://github.com/pytorch/pytorch/pull/18006
ghimport-source-id: e267ece1ac03e0d17e01dddf4a77f52421859435
Stack:
* **#18006 Fix lint in test_nn.py**
Signed-off-by: Edward Z. Yang <ezyang@fb.com>
Reviewed By: eellison
Differential Revision: D14458108
fbshipit-source-id: 18ee6199447efed55a922cff5b3ad940a21c0536
Diffstat (limited to 'test/test_nn.py')
-rw-r--r-- | test/test_nn.py | 20 |
1 files changed, 10 insertions, 10 deletions
diff --git a/test/test_nn.py b/test/test_nn.py index a1541fc3ae..eafeddbe28 100644 --- a/test/test_nn.py +++ b/test/test_nn.py @@ -1886,7 +1886,7 @@ class TestNN(NNTestCase): self.assertEqual(wrapped_m(input), pre_remove_out) m = torch.nn.utils.spectral_norm(m) - for i in range(3): + for _ in range(3): pre_remove_out = wrapped_m(input) m = torch.nn.utils.remove_spectral_norm(m) self.assertEqual(wrapped_m(input), pre_remove_out) @@ -2215,7 +2215,7 @@ class TestNN(NNTestCase): probs = logits.softmax(dim=-1) counts = torch.zeros_like(logits) - for draw in range(num_draws): + for _ in range(num_draws): y_draw = F.gumbel_softmax(logits, hard=True) counts = counts + y_draw @@ -4022,7 +4022,7 @@ class TestNN(NNTestCase): my_stream = torch.cuda.Stream() results[t] = input with torch.cuda.stream(my_stream): - for i in range(test_iters): + for _ in range(test_iters): # If all threads are sharing the same cudnn handle, # the following sequence may occur: # thread 0 calls setCuDNNStreamToCurrent() @@ -4035,7 +4035,7 @@ class TestNN(NNTestCase): results[t].div_(4.0) torch.cuda.current_stream().wait_stream(my_stream) - for trial in range(trials): + for _ in range(trials): for t in range(num_threads): results[t] = torch.ones((1, 1, 2048, 2048), device='cuda') @@ -4320,7 +4320,7 @@ class TestNN(NNTestCase): input = torch.randn(3, 10) hx = torch.randn(3, 20) cell = module(10, 20, bias=bias) - for i in range(6): + for _ in range(6): hx = cell(input, hx) hx.sum().backward() @@ -4337,7 +4337,7 @@ class TestNN(NNTestCase): input = Variable(cast(torch.randn(3, 5))) target = Variable(cast(torch.randn(5, 3))) - for name, fn in losses.items(): + for _name, fn in losses.items(): self.assertRaises(Exception, lambda: fn(input, target)) def test_loss_equal_input_target_shape(self): @@ -4770,7 +4770,7 @@ class TestNN(NNTestCase): hx = torch.randn(3, 20) cx = torch.randn(3, 20) lstm = nn.LSTMCell(10, 20, bias=bias) - for i in range(6): + for _ in range(6): hx, cx = lstm(input, (hx, cx)) (hx + cx).sum().backward() @@ -4815,7 +4815,7 @@ class TestNN(NNTestCase): with torch.no_grad(): weight.set_(weight_data) - for i in range(2): + for _ in range(2): with warnings.catch_warnings(record=True) as w: output_noncontig = rnn(input, hx) if first_warn: @@ -5057,7 +5057,7 @@ class TestNN(NNTestCase): output = rnn(input) output[0].sum().backward(retain_graph=True) grads = [input.grad.data.clone()] + [p.grad.data.clone() for p in rnn.parameters()] - for i in range(4): + for _ in range(4): rnn.zero_grad() input.grad.data.zero_() output[0].sum().backward(retain_graph=True) @@ -5932,7 +5932,7 @@ class TestNN(NNTestCase): } input = torch.randn(2, 1, requires_grad=True) - for name, fn in losses.items(): + for _name, fn in losses.items(): for requires_grad in [True, False]: # When target.requires_grad=True, its impl is in Python, while the other is in TH. target = torch.randn(2, 10, requires_grad=requires_grad) |