From f090e4a99cb834060ff59655667f8e9930e3ded3 Mon Sep 17 00:00:00 2001 From: x54-729 <17307130121@fudan.edu.cn> Date: Wed, 28 Sep 2022 13:47:26 +0800 Subject: [PATCH] =?UTF-8?q?.Jenkinsfile=20=E6=B7=BB=E5=8A=A0=20shm-size=20?= =?UTF-8?q?=E8=AE=BE=E7=BD=AE=EF=BC=9Btest=5Fseq2seq=5Fmodel=20=E8=B0=83?= =?UTF-8?q?=E6=95=B4=E8=AE=AD=E7=BB=83=E5=8F=82=E6=95=B0?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .Jenkinsfile | 6 +++--- tests/models/torch/test_seq2seq_model.py | 4 ++-- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.Jenkinsfile b/.Jenkinsfile index 75156959..f2740e5b 100644 --- a/.Jenkinsfile +++ b/.Jenkinsfile @@ -46,7 +46,7 @@ pipeline { agent { docker { image 'fnlp:torch-1.6' - args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all' + args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all --shm-size 1G' } } steps { @@ -62,7 +62,7 @@ pipeline { agent { docker { image 'fnlp:paddle' - args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all' + args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all --shm-size 1G' } } steps { @@ -82,7 +82,7 @@ pipeline { // agent { // docker { // image 'fnlp:jittor' - // args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all' + // args '-u root:root -v ${JENKINS_HOME}/html/docs:/docs -v ${JENKINS_HOME}/html/_ci:/ci --gpus all --shm-size 1G' // } // } // steps { diff --git a/tests/models/torch/test_seq2seq_model.py b/tests/models/torch/test_seq2seq_model.py index ee1775d3..abae1f67 100755 --- a/tests/models/torch/test_seq2seq_model.py +++ b/tests/models/torch/test_seq2seq_model.py @@ -25,11 +25,11 @@ def prepare_env(): def train_model(model, src_words_idx, tgt_words_idx, tgt_seq_len, src_seq_len): - optimizer = optim.Adam(model.parameters(), lr=1e-2) + optimizer = optim.Adam(model.parameters(), lr=5e-3) mask = seq_len_to_mask(tgt_seq_len).eq(0) target = tgt_words_idx.masked_fill(mask, -100) - for i in range(100): + for i in range(50): optimizer.zero_grad() pred = model(src_words_idx, tgt_words_idx, src_seq_len)['pred'] # bsz x max_len x vocab_size loss = F.cross_entropy(pred.transpose(1, 2), target)