Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
|
|
4 years ago | |
|---|---|---|
| .. | ||
| models | 4 years ago | |
| parallel | 4 years ago | |
| README.md | 4 years ago | |
| local_allreduce.yml | 4 years ago | |
| local_ps.yml | 4 years ago | |
| remote_allreduce.yml | 4 years ago | |
| remote_ps.yml | 4 years ago | |
| run_mlp.py | 4 years ago | |
| run_wdl.py | 4 years ago | |
This directory contains examples using heturun command.
# Local Data Parallel Using AllReduce
heturun -c local_allreduce.yml python run_mlp.py --config lar
# Local Data Parallel Using AllReduce for Dense Parameters and PS for Sparse(Embedding) Parameters
heturun -c local_ps.yml python run_wdl.py --config lhy
# Local Data Parallel Using PS
heturun -c local_ps.yml python run_mlp.py --config lps
heturun -c local_ps.yml python run_wdl.py --config lps
# Distributed Data Parallel Using AllReduce
heturun -c remote_allreduce.yml python run_mlp.py --config rar
# Distributed Data Parallel Using AllReduce for Dense Parameters and PS for Sparse(Embedding) Parameters
heturun -c remote_ps.yml python run_wdl.py --config rhy
# Distributed Data Parallel Using PS
heturun -c remote_ps.yml python run_mlp.py --config rps
heturun -c remote_ps.yml python run_wdl.py --config rps
parallel directory.分布式深度学习系统
Python C++ Cuda C CMake other