Are you sure you want to delete this task? Once this task is deleted, it cannot be recovered.
|
|
5 years ago | |
|---|---|---|
| .. | ||
| src | 5 years ago | |
| README.md | 5 years ago | |
| run_multinpu_train.sh | 5 years ago | |
| test.py | 5 years ago | |
| train.py | 5 years ago | |
| train_and_test.py | 5 years ago | |
| train_and_test_multinpu.py | 5 years ago | |
recommendation Model
This is an implementation of WideDeep as described in the Wide & Deep Learning for Recommender System paper.
WideDeep model jointly trained wide linear models and deep neural network, which combined the benefits of memorization and generalization for recommender systems.
The Criteo datasets are used for model training and evaluation.
To download the dataset, please install Pandas package first. Then issue the following command:
bash download.sh
The entire code structure is as following:
|--- wide_and_deep/
train_and_test.py "Entrance of Wide&Deep model training and evaluation"
test.py "Entrance of Wide&Deep model evaluation"
train.py "Entrance of Wide&Deep model training"
train_and_test_multinpu.py "Entrance of Wide&Deep model data parallel training and evaluation"
|--- src/ "entrance of training and evaluation"
config.py "parameters configuration"
dataset.py "Dataset loader class"
WideDeep.py "Model structure"
callbacks.py "Callback class for training and evaluation"
metrics.py "Metric class"
To train and evaluate the model, issue the following command:
python train_and_test.py
Arguments:
--data_path: This should be set to the same directory given to the data_download's data_dir argument.--epochs: Total train epochs.--batch_size: Training batch size.--eval_batch_size: Eval batch size.--field_size: The number of features.--vocab_size: The total features of dataset.--emb_dim: The dense embedding dimension of sparse feature.--deep_layers_dim: The dimension of all deep layers.--deep_layers_act: The activation of all deep layers.--keep_prob: The rate to keep in dropout layer.--ckpt_path:The location of the checkpoint file.--eval_file_name : Eval output file.--loss_file_name : Loss output file.To train the model, issue the following command:
python train.py
Arguments:
--data_path: This should be set to the same directory given to the data_download's data_dir argument.--epochs: Total train epochs.--batch_size: Training batch size.--eval_batch_size: Eval batch size.--field_size: The number of features.--vocab_size: The total features of dataset.--emb_dim: The dense embedding dimension of sparse feature.--deep_layers_dim: The dimension of all deep layers.--deep_layers_act: The activation of all deep layers.--keep_prob: The rate to keep in dropout layer.--ckpt_path:The location of the checkpoint file.--eval_file_name : Eval output file.--loss_file_name : Loss output file.To evaluate the model, issue the following command:
python test.py
Arguments:
--data_path: This should be set to the same directory given to the data_download's data_dir argument.--epochs: Total train epochs.--batch_size: Training batch size.--eval_batch_size: Eval batch size.--field_size: The number of features.--vocab_size: The total features of dataset.--emb_dim: The dense embedding dimension of sparse feature.--deep_layers_dim: The dimension of all deep layers.--deep_layers_act: The activation of all deep layers.--keep_prob: The rate to keep in dropout layer.--ckpt_path:The location of the checkpoint file.--eval_file_name : Eval output file.--loss_file_name : Loss output file.There are other arguments about models and training process. Use the --help or -h flag to get a full list of possible arguments with detailed descriptions.
MindSpore is a new open source deep learning training/inference framework that could be used for mobile, edge and cloud scenarios.
C++ Python Text Unity3D Asset C other