From cf0b2c2d35f9ac7cdbc13eaa30cef80e000f5bfb Mon Sep 17 00:00:00 2001 From: yunfan Date: Fri, 9 Nov 2018 18:22:24 +0800 Subject: [PATCH] update trainer --- fastNLP/core/trainer.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/fastNLP/core/trainer.py b/fastNLP/core/trainer.py index 23f6fecc..d1881297 100644 --- a/fastNLP/core/trainer.py +++ b/fastNLP/core/trainer.py @@ -171,11 +171,11 @@ class Trainer(object): loss = self.get_loss(prediction, batch_y) self.grad_backward(loss) - if torch.rand(1).item() < 0.001: - print('[grads at epoch: {:>3} step: {:>4}]'.format(kwargs['epoch'], step)) - for name, p in self._model.named_parameters(): - if p.requires_grad: - print('\t{} {} {}'.format(name, tuple(p.size()), torch.sum(p.grad).item())) + # if torch.rand(1).item() < 0.001: + # print('[grads at epoch: {:>3} step: {:>4}]'.format(kwargs['epoch'], step)) + # for name, p in self._model.named_parameters(): + # if p.requires_grad: + # print('\t{} {} {}'.format(name, tuple(p.size()), torch.sum(p.grad).item())) self.update() self._summary_writer.add_scalar("loss", loss.item(), global_step=step)