Browse Source

!571 optimize mindwizard network templates

Merge pull request !571 from liangyongxiong/wizard
tags/v1.0.0
mindspore-ci-bot Gitee 5 years ago
parent
commit
3cfb8e7290
24 changed files with 9 additions and 25 deletions
  1. +4
    -4
      mindinsight/wizard/README_CN.md
  2. +1
    -0
      mindinsight/wizard/base/source_file.py
  3. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train.sh-tpl
  4. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train_gpu.sh-tpl
  5. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval.sh-tpl
  6. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval_gpu.sh-tpl
  7. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train.sh-tpl
  8. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train_gpu.sh-tpl
  9. +1
    -1
      mindinsight/wizard/conf/templates/network/alexnet/src/config.py-tpl
  10. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train.sh-tpl
  11. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train_gpu.sh-tpl
  12. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval.sh-tpl
  13. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval_gpu.sh-tpl
  14. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train.sh-tpl
  15. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train_gpu.sh-tpl
  16. +1
    -1
      mindinsight/wizard/conf/templates/network/lenet/src/config.py-tpl
  17. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train.sh-tpl
  18. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train_gpu.sh-tpl
  19. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval.sh-tpl
  20. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval_gpu.sh-tpl
  21. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train.sh-tpl
  22. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train_gpu.sh-tpl
  23. +1
    -1
      mindinsight/wizard/conf/templates/network/resnet50/src/config.py-tpl
  24. +1
    -0
      mindinsight/wizard/create_project.py

+ 4
- 4
mindinsight/wizard/README_CN.md View File

@@ -4,7 +4,7 @@

## 介绍

MindWizard是一款快速生成经典网络脚本的工具。工具根据用户选择,组合模型、超参、数据集等网络参数,自动生成目标网络脚本,生成的网络脚本可以在Ascend或GPU等环境上进行训练和推理
MindWizard是一款快速生成经典网络脚本的工具。工具根据用户选择,组合模型、超参、数据集等网络参数,自动生成目标网络脚本,生成的网络脚本可以在Ascend或GPU等环境上进行训练和评估

## 安装

@@ -33,7 +33,7 @@ optional arguments:

2. 请选择数据集(MNIST / Cifar10 / ImageNet / ...)

生成脚本后,用户可执行训练和推理,详细介绍可参考网络脚本工程中的README。
生成脚本后,用户可执行训练和评估,详细介绍可参考网络脚本工程中的README。

## 网络脚本工程结构

@@ -42,14 +42,14 @@ project
|- script
| |- run_standalone_train.sh # 单卡训练脚本
| |- run_distribute_train.sh # 多卡训练脚本
| |- run_eval.sh # 推理脚本
| |- run_eval.sh # 评估脚本
| |- ...
|- src
| |- config.py # 参数配置
| |- dataset.py # 数据集处理
| |- lenet.py/resent.py/... # 网络定义
| |- ...
|- eval.py # 网络推理
|- eval.py # 网络评估
|- train.py # 网络训练
|- README.md
```


+ 1
- 0
mindinsight/wizard/base/source_file.py View File

@@ -45,6 +45,7 @@ class SourceFile:
fp.write(self.content)
try:
shutil.copymode(self.template_file_path, new_file_path)
os.chmod(new_file_path, stat.S_IRUSR | stat.S_IWUSR)
self.set_writeable(new_file_path)
if new_file_path.endswith('.sh'):
self.set_executable(new_file_path)


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train.sh-tpl View File

@@ -67,7 +67,6 @@ do
rm -rf ./train_parallel$i
mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -54,7 +54,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel
mkdir ./train_parallel
cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel
cd ./train_parallel || exit
echo "start training"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit
echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit
echo "start training"


+ 1
- 1
mindinsight/wizard/conf/templates/network/alexnet/src/config.py-tpl View File

@@ -28,7 +28,7 @@ cfg = edict({
'lr': 0.002,
"momentum": 0.9,
{% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %}
'lr': 0.001,
{% endif %}


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train.sh-tpl View File

@@ -68,7 +68,6 @@ do
rm -rf ./train_parallel$i
mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -53,7 +53,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel
mkdir ./train_parallel
cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel
cd ./train_parallel || exit



+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train.sh-tpl View File

@@ -60,7 +60,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit
echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -61,7 +61,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit



+ 1
- 1
mindinsight/wizard/conf/templates/network/lenet/src/config.py-tpl View File

@@ -23,7 +23,7 @@ cfg = edict({
'lr': 0.01,
"momentum": 0.9,
{% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %}
'lr': 0.001,
{% endif %}


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train.sh-tpl View File

@@ -67,7 +67,6 @@ do
rm -rf ./train_parallel$i
mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -54,7 +54,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel
mkdir ./train_parallel
cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel
cd ./train_parallel || exit



+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi
mkdir ./eval
cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval
cd ./eval || exit
env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit
echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi
mkdir ./train
cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train
cd ./train || exit
echo "start training"


+ 1
- 1
mindinsight/wizard/conf/templates/network/resnet50/src/config.py-tpl View File

@@ -30,7 +30,7 @@ cfg = ed({
"momentum": 0.9,
"lr": 0.01,
{% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %}
'lr': 0.001,
{% endif %}


+ 1
- 0
mindinsight/wizard/create_project.py View File

@@ -91,6 +91,7 @@ class CreateProject(BaseCommand):
if not choice:
click.secho(textwrap.dedent("Network is required."), fg='red')

click.secho(textwrap.dedent("Your choice is %s." % network_type_choices[choice - 1]), fg='yellow')
return network_type_choices[choice - 1]

@staticmethod


Loading…
Cancel
Save