diff --git a/modelscope/pipelines/nlp/distributed_plug_pipeline.py b/modelscope/pipelines/nlp/distributed_plug_pipeline.py index fe42e472..fc80fe38 100644 --- a/modelscope/pipelines/nlp/distributed_plug_pipeline.py +++ b/modelscope/pipelines/nlp/distributed_plug_pipeline.py @@ -65,8 +65,7 @@ class DistributedPlugPipeline(DistributedPipeline): sequence_length=sequence_length, **kwargs) super().__init__(model, preprocessor=preprocessor, **kwargs) - assert hasattr(preprocessor, 'tokenizer') - self.cls_token_id = preprocessor.tokenizer.cls_token_id + self.cls_token_id = preprocessor.nlp_tokenizer.tokenizer.cls_token_id @classmethod def _forward_one(cls, inputs: Dict[str, Any]) -> Dict[str, Any]: @@ -105,6 +104,6 @@ class DistributedPlugPipeline(DistributedPipeline): from modelscope.outputs import OutputKeys generate_context = inputs['generate_context'] generate_context = ''.join( - self.preprocessor.tokenizer.convert_ids_to_tokens( + self.preprocessor.nlp_tokenizer.tokenizer.convert_ids_to_tokens( generate_context)).replace('[UNK]', '“').replace('##', '') return {OutputKeys.TEXT: generate_context} diff --git a/modelscope/trainers/nlp/plug_trainer.py b/modelscope/trainers/nlp/plug_trainer.py index 6d0a0c01..7d7d830c 100644 --- a/modelscope/trainers/nlp/plug_trainer.py +++ b/modelscope/trainers/nlp/plug_trainer.py @@ -66,9 +66,9 @@ class PlugTrainer(NlpEpochBasedTrainer): from deepspeed.ops.adam import DeepSpeedCPUAdam model = self.model - embeddings = model.module.module.model.bert.embeddings - layers = model.module.module.model.bert.encoder.layer - dec_layers = model.module.module.model.decoder.decoder + embeddings = model.module.model.bert.embeddings + layers = model.module.model.bert.encoder.layer + dec_layers = model.module.model.decoder.decoder param_groups = [] param_groups += list( self._get_params_for_weight_decay_optimization(layers)) @@ -160,7 +160,7 @@ class PlugTrainer(NlpEpochBasedTrainer): def evaluation_step(self, data): # wapper 1: DeepspeedEngine, wapper 2: DDP - model = self.model.module.module + model = self.model.module model.eval() # model: fp16 wapper; model.module : distributedPlug