Browse Source

optimize network templates

tags/v1.0.0
liangyongxiong 5 years ago
parent
commit
2c19c9b8c9
24 changed files with 9 additions and 25 deletions
  1. +4
    -4
      mindinsight/wizard/README_CN.md
  2. +1
    -0
      mindinsight/wizard/base/source_file.py
  3. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train.sh-tpl
  4. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train_gpu.sh-tpl
  5. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval.sh-tpl
  6. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval_gpu.sh-tpl
  7. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train.sh-tpl
  8. +0
    -1
      mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train_gpu.sh-tpl
  9. +1
    -1
      mindinsight/wizard/conf/templates/network/alexnet/src/config.py-tpl
  10. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train.sh-tpl
  11. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train_gpu.sh-tpl
  12. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval.sh-tpl
  13. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval_gpu.sh-tpl
  14. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train.sh-tpl
  15. +0
    -1
      mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train_gpu.sh-tpl
  16. +1
    -1
      mindinsight/wizard/conf/templates/network/lenet/src/config.py-tpl
  17. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train.sh-tpl
  18. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train_gpu.sh-tpl
  19. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval.sh-tpl
  20. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval_gpu.sh-tpl
  21. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train.sh-tpl
  22. +0
    -1
      mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train_gpu.sh-tpl
  23. +1
    -1
      mindinsight/wizard/conf/templates/network/resnet50/src/config.py-tpl
  24. +1
    -0
      mindinsight/wizard/create_project.py

+ 4
- 4
mindinsight/wizard/README_CN.md View File

@@ -4,7 +4,7 @@


## 介绍 ## 介绍


MindWizard是一款快速生成经典网络脚本的工具。工具根据用户选择,组合模型、超参、数据集等网络参数,自动生成目标网络脚本,生成的网络脚本可以在Ascend或GPU等环境上进行训练和推理
MindWizard是一款快速生成经典网络脚本的工具。工具根据用户选择,组合模型、超参、数据集等网络参数,自动生成目标网络脚本,生成的网络脚本可以在Ascend或GPU等环境上进行训练和评估


## 安装 ## 安装


@@ -33,7 +33,7 @@ optional arguments:


2. 请选择数据集(MNIST / Cifar10 / ImageNet / ...) 2. 请选择数据集(MNIST / Cifar10 / ImageNet / ...)


生成脚本后,用户可执行训练和推理,详细介绍可参考网络脚本工程中的README。
生成脚本后,用户可执行训练和评估,详细介绍可参考网络脚本工程中的README。


## 网络脚本工程结构 ## 网络脚本工程结构


@@ -42,14 +42,14 @@ project
|- script |- script
| |- run_standalone_train.sh # 单卡训练脚本 | |- run_standalone_train.sh # 单卡训练脚本
| |- run_distribute_train.sh # 多卡训练脚本 | |- run_distribute_train.sh # 多卡训练脚本
| |- run_eval.sh # 推理脚本
| |- run_eval.sh # 评估脚本
| |- ... | |- ...
|- src |- src
| |- config.py # 参数配置 | |- config.py # 参数配置
| |- dataset.py # 数据集处理 | |- dataset.py # 数据集处理
| |- lenet.py/resent.py/... # 网络定义 | |- lenet.py/resent.py/... # 网络定义
| |- ... | |- ...
|- eval.py # 网络推理
|- eval.py # 网络评估
|- train.py # 网络训练 |- train.py # 网络训练
|- README.md |- README.md
``` ```


+ 1
- 0
mindinsight/wizard/base/source_file.py View File

@@ -45,6 +45,7 @@ class SourceFile:
fp.write(self.content) fp.write(self.content)
try: try:
shutil.copymode(self.template_file_path, new_file_path) shutil.copymode(self.template_file_path, new_file_path)
os.chmod(new_file_path, stat.S_IRUSR | stat.S_IWUSR)
self.set_writeable(new_file_path) self.set_writeable(new_file_path)
if new_file_path.endswith('.sh'): if new_file_path.endswith('.sh'):
self.set_executable(new_file_path) self.set_executable(new_file_path)


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train.sh-tpl View File

@@ -67,7 +67,6 @@ do
rm -rf ./train_parallel$i rm -rf ./train_parallel$i
mkdir ./train_parallel$i mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID" echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -54,7 +54,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel rm -rf ./train_parallel
mkdir ./train_parallel mkdir ./train_parallel
cp ../*.py ./train_parallel cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel cp -r ../src ./train_parallel
cd ./train_parallel || exit cd ./train_parallel || exit
echo "start training" echo "start training"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit
echo "start training for device $DEVICE_ID" echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/alexnet/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit
echo "start training" echo "start training"


+ 1
- 1
mindinsight/wizard/conf/templates/network/alexnet/src/config.py-tpl View File

@@ -28,7 +28,7 @@ cfg = edict({
'lr': 0.002, 'lr': 0.002,
"momentum": 0.9, "momentum": 0.9,
{% elif optimizer=='SGD' %} {% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %} {% else %}
'lr': 0.001, 'lr': 0.001,
{% endif %} {% endif %}


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train.sh-tpl View File

@@ -68,7 +68,6 @@ do
rm -rf ./train_parallel$i rm -rf ./train_parallel$i
mkdir ./train_parallel$i mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID" echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -53,7 +53,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel rm -rf ./train_parallel
mkdir ./train_parallel mkdir ./train_parallel
cp ../*.py ./train_parallel cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel cp -r ../src ./train_parallel
cd ./train_parallel || exit cd ./train_parallel || exit




+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train.sh-tpl View File

@@ -60,7 +60,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit
echo "start training for device $DEVICE_ID" echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/lenet/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -61,7 +61,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit




+ 1
- 1
mindinsight/wizard/conf/templates/network/lenet/src/config.py-tpl View File

@@ -23,7 +23,7 @@ cfg = edict({
'lr': 0.01, 'lr': 0.01,
"momentum": 0.9, "momentum": 0.9,
{% elif optimizer=='SGD' %} {% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %} {% else %}
'lr': 0.001, 'lr': 0.001,
{% endif %} {% endif %}


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train.sh-tpl View File

@@ -67,7 +67,6 @@ do
rm -rf ./train_parallel$i rm -rf ./train_parallel$i
mkdir ./train_parallel$i mkdir ./train_parallel$i
cp ../*.py ./train_parallel$i cp ../*.py ./train_parallel$i
cp *.sh ./train_parallel$i
cp -r ../src ./train_parallel$i cp -r ../src ./train_parallel$i
cd ./train_parallel$i || exit cd ./train_parallel$i || exit
echo "start training for rank $RANK_ID, device $DEVICE_ID" echo "start training for rank $RANK_ID, device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_distribute_train_gpu.sh-tpl View File

@@ -54,7 +54,6 @@ export RANK_SIZE=$DEVICE_NUM
rm -rf ./train_parallel rm -rf ./train_parallel
mkdir ./train_parallel mkdir ./train_parallel
cp ../*.py ./train_parallel cp ../*.py ./train_parallel
cp *.sh ./train_parallel
cp -r ../src ./train_parallel cp -r ../src ./train_parallel
cd ./train_parallel || exit cd ./train_parallel || exit




+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval.sh-tpl View File

@@ -56,7 +56,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_eval_gpu.sh-tpl View File

@@ -57,7 +57,6 @@ then
fi fi
mkdir ./eval mkdir ./eval
cp ../*.py ./eval cp ../*.py ./eval
cp *.sh ./eval
cp -r ../src ./eval cp -r ../src ./eval
cd ./eval || exit cd ./eval || exit
env > env.log env > env.log


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit
echo "start training for device $DEVICE_ID" echo "start training for device $DEVICE_ID"


+ 0
- 1
mindinsight/wizard/conf/templates/network/resnet50/scripts/run_standalone_train_gpu.sh-tpl View File

@@ -59,7 +59,6 @@ then
fi fi
mkdir ./train mkdir ./train
cp ../*.py ./train cp ../*.py ./train
cp *.sh ./train
cp -r ../src ./train cp -r ../src ./train
cd ./train || exit cd ./train || exit
echo "start training" echo "start training"


+ 1
- 1
mindinsight/wizard/conf/templates/network/resnet50/src/config.py-tpl View File

@@ -30,7 +30,7 @@ cfg = ed({
"momentum": 0.9, "momentum": 0.9,
"lr": 0.01, "lr": 0.01,
{% elif optimizer=='SGD' %} {% elif optimizer=='SGD' %}
'lr': 0.1,
'lr': 0.01,
{% else %} {% else %}
'lr': 0.001, 'lr': 0.001,
{% endif %} {% endif %}


+ 1
- 0
mindinsight/wizard/create_project.py View File

@@ -91,6 +91,7 @@ class CreateProject(BaseCommand):
if not choice: if not choice:
click.secho(textwrap.dedent("Network is required."), fg='red') click.secho(textwrap.dedent("Network is required."), fg='red')


click.secho(textwrap.dedent("Your choice is %s." % network_type_choices[choice - 1]), fg='yellow')
return network_type_choices[choice - 1] return network_type_choices[choice - 1]


@staticmethod @staticmethod


Loading…
Cancel
Save