add feature: check and release stream and event resources when lazy recomplie mode refactor: move data_dumper.cc to common/dumppull/2082/head
| @@ -124,7 +124,7 @@ set(EXECUTOR_SRC_LIST | |||||
| "graph/load/graph_loader.cc" | "graph/load/graph_loader.cc" | ||||
| "graph/load/model_manager/aipp_utils.cc" | "graph/load/model_manager/aipp_utils.cc" | ||||
| "graph/load/model_manager/cpu_queue_schedule.cc" | "graph/load/model_manager/cpu_queue_schedule.cc" | ||||
| "graph/load/model_manager/data_dumper.cc" | |||||
| "common/dump/data_dumper.cc" | |||||
| "graph/load/model_manager/data_inputer.cc" | "graph/load/model_manager/data_inputer.cc" | ||||
| "graph/load/model_manager/davinci_model.cc" | "graph/load/model_manager/davinci_model.cc" | ||||
| "graph/load/model_manager/model_manager.cc" | "graph/load/model_manager/model_manager.cc" | ||||
| @@ -14,7 +14,7 @@ | |||||
| * limitations under the License. | * limitations under the License. | ||||
| */ | */ | ||||
| #include "graph/load/model_manager/data_dumper.h" | |||||
| #include "data_dumper.h" | |||||
| #include <cstdlib> | #include <cstdlib> | ||||
| #include <ctime> | #include <ctime> | ||||
| @@ -20,7 +20,7 @@ | |||||
| #include <map> | #include <map> | ||||
| #include "framework/common/debug/ge_log.h" | #include "framework/common/debug/ge_log.h" | ||||
| #include "framework/common/debug/log.h" | #include "framework/common/debug/log.h" | ||||
| #include "graph/load/model_manager/data_dumper.h" | |||||
| #include "data_dumper.h" | |||||
| namespace ge { | namespace ge { | ||||
| class OpdebugRegister { | class OpdebugRegister { | ||||
| @@ -25,7 +25,7 @@ set(SRC_LIST | |||||
| "../graph/load/model_manager/model_utils.cc" | "../graph/load/model_manager/model_utils.cc" | ||||
| "../graph/load/model_manager/aipp_utils.cc" | "../graph/load/model_manager/aipp_utils.cc" | ||||
| "../graph/load/model_manager/data_inputer.cc" | "../graph/load/model_manager/data_inputer.cc" | ||||
| "../graph/load/model_manager/data_dumper.cc" | |||||
| "../common/dump/data_dumper.cc" | |||||
| "../graph/load/model_manager/zero_copy_task.cc" | "../graph/load/model_manager/zero_copy_task.cc" | ||||
| "../graph/load/model_manager/zero_copy_offset.cc" | "../graph/load/model_manager/zero_copy_offset.cc" | ||||
| "../graph/load/model_manager/task_info/task_info.cc" | "../graph/load/model_manager/task_info/task_info.cc" | ||||
| @@ -6,6 +6,7 @@ local_ge_executor_src_files := \ | |||||
| ../common/dump/dump_properties.cc \ | ../common/dump/dump_properties.cc \ | ||||
| ../common/dump/dump_manager.cc \ | ../common/dump/dump_manager.cc \ | ||||
| ../common/dump/dump_op.cc \ | ../common/dump/dump_op.cc \ | ||||
| ../common/dump/data_dumper.cc \ | |||||
| ../common/ge/plugin_manager.cc \ | ../common/ge/plugin_manager.cc \ | ||||
| ../common/ge/op_tiling_manager.cc \ | ../common/ge/op_tiling_manager.cc \ | ||||
| ../common/profiling/ge_profiling.cc \ | ../common/profiling/ge_profiling.cc \ | ||||
| @@ -29,7 +30,6 @@ local_ge_executor_src_files := \ | |||||
| ../graph/load/model_manager/model_utils.cc \ | ../graph/load/model_manager/model_utils.cc \ | ||||
| ../graph/load/model_manager/aipp_utils.cc \ | ../graph/load/model_manager/aipp_utils.cc \ | ||||
| ../graph/load/model_manager/data_inputer.cc \ | ../graph/load/model_manager/data_inputer.cc \ | ||||
| ../graph/load/model_manager/data_dumper.cc \ | |||||
| ../graph/load/model_manager/zero_copy_task.cc \ | ../graph/load/model_manager/zero_copy_task.cc \ | ||||
| ../graph/load/model_manager/zero_copy_offset.cc \ | ../graph/load/model_manager/zero_copy_offset.cc \ | ||||
| ../graph/load/model_manager/task_info/task_info.cc \ | ../graph/load/model_manager/task_info/task_info.cc \ | ||||
| @@ -239,7 +239,7 @@ OME_HOST_SRC_FILES := \ | |||||
| graph/load/model_manager/cpu_queue_schedule.cc \ | graph/load/model_manager/cpu_queue_schedule.cc \ | ||||
| graph/load/model_manager/zero_copy_task.cc \ | graph/load/model_manager/zero_copy_task.cc \ | ||||
| graph/load/model_manager/zero_copy_offset.cc \ | graph/load/model_manager/zero_copy_offset.cc \ | ||||
| graph/load/model_manager/data_dumper.cc \ | |||||
| common/dump/data_dumper.cc \ | |||||
| graph/load/model_manager/task_info/task_info.cc \ | graph/load/model_manager/task_info/task_info.cc \ | ||||
| graph/load/model_manager/task_info/event_record_task_info.cc \ | graph/load/model_manager/task_info/event_record_task_info.cc \ | ||||
| graph/load/model_manager/task_info/event_wait_task_info.cc \ | graph/load/model_manager/task_info/event_wait_task_info.cc \ | ||||
| @@ -29,6 +29,7 @@ LIBGE_LOCAL_SRC_FILES := \ | |||||
| common/dump/dump_manager.cc \ | common/dump/dump_manager.cc \ | ||||
| common/dump/dump_properties.cc \ | common/dump/dump_properties.cc \ | ||||
| common/dump/dump_op.cc \ | common/dump/dump_op.cc \ | ||||
| common/dump/data_dumper.cc \ | |||||
| common/profiling/ge_profiling.cc \ | common/profiling/ge_profiling.cc \ | ||||
| common/profiling/ge_runner_profiling.cc \ | common/profiling/ge_runner_profiling.cc \ | ||||
| engine_manager/dnnengine_manager.cc \ | engine_manager/dnnengine_manager.cc \ | ||||
| @@ -55,7 +56,6 @@ LIBGE_LOCAL_SRC_FILES := \ | |||||
| graph/label/while_label_maker.cc \ | graph/label/while_label_maker.cc \ | ||||
| graph/load/graph_loader.cc \ | graph/load/graph_loader.cc \ | ||||
| graph/load/model_manager/cpu_queue_schedule.cc \ | graph/load/model_manager/cpu_queue_schedule.cc \ | ||||
| graph/load/model_manager/data_dumper.cc \ | |||||
| graph/load/model_manager/data_inputer.cc \ | graph/load/model_manager/data_inputer.cc \ | ||||
| graph/load/model_manager/davinci_model.cc \ | graph/load/model_manager/davinci_model.cc \ | ||||
| graph/load/model_manager/davinci_model_parser.cc \ | graph/load/model_manager/davinci_model_parser.cc \ | ||||
| @@ -35,7 +35,7 @@ | |||||
| #include "framework/common/util.h" | #include "framework/common/util.h" | ||||
| #include "graph/debug/ge_attr_define.h" | #include "graph/debug/ge_attr_define.h" | ||||
| #include "graph/load/model_manager/aipp_utils.h" | #include "graph/load/model_manager/aipp_utils.h" | ||||
| #include "graph/load/model_manager/data_dumper.h" | |||||
| #include "common/dump/data_dumper.h" | |||||
| #include "graph/load/model_manager/data_inputer.h" | #include "graph/load/model_manager/data_inputer.h" | ||||
| #include "graph/load/model_manager/model_utils.h" | #include "graph/load/model_manager/model_utils.h" | ||||
| #include "graph/load/model_manager/zero_copy_offset.h" | #include "graph/load/model_manager/zero_copy_offset.h" | ||||
| @@ -288,6 +288,8 @@ class DavinciModel { | |||||
| const vector<rtLabel_t> &GetLabelList() const { return label_list_; } | const vector<rtLabel_t> &GetLabelList() const { return label_list_; } | ||||
| uint64_t GetAllStreamNum() const { return stream_list_.size() + all_hccl_stream_list_.size(); } | |||||
| Status GetLabelGotoAddr(uint32_t label_index, rtMemType_t memory_type, void *&addr, uint32_t &size); | Status GetLabelGotoAddr(uint32_t label_index, rtMemType_t memory_type, void *&addr, uint32_t &size); | ||||
| Status DestroyThread(); | Status DestroyThread(); | ||||
| @@ -44,9 +44,13 @@ const std::string kCmdTypeProfModelSubscribe = "prof_model_subscribe"; | |||||
| const std::string kCmdTypeProfModelUnsubscribe = "prof_model_cancel_subscribe"; | const std::string kCmdTypeProfModelUnsubscribe = "prof_model_cancel_subscribe"; | ||||
| const char *const kBatchLoadBuf = "batchLoadsoFrombuf"; | const char *const kBatchLoadBuf = "batchLoadsoFrombuf"; | ||||
| const char *const kDeleteCustOp = "deleteCustOp"; | const char *const kDeleteCustOp = "deleteCustOp"; | ||||
| const char *const kStreamResource = "stream"; | |||||
| const char *const kEventResource = "event"; | |||||
| const char *const kIsCopyOutputAddr = "1"; | |||||
| const int kTimeSpecNano = 1000000000; | const int kTimeSpecNano = 1000000000; | ||||
| const int kTimeSpecMiro = 1000000; | const int kTimeSpecMiro = 1000000; | ||||
| const int kOpNameMaxSize = 100; | const int kOpNameMaxSize = 100; | ||||
| const int kMaxEventNum = 1024; | |||||
| const uint64_t kInferSessionId = 0; | const uint64_t kInferSessionId = 0; | ||||
| #pragma pack(push, 1) | #pragma pack(push, 1) | ||||
| struct CustAicpuSoBuf { | struct CustAicpuSoBuf { | ||||
| @@ -360,6 +364,26 @@ Status ModelManager::LoadModelOnline(uint32_t &model_id, const shared_ptr<ge::Ge | |||||
| davinci_model->SetDumpProperties(dump_properties); | davinci_model->SetDumpProperties(dump_properties); | ||||
| dump_properties_ = dump_properties; | dump_properties_ = dump_properties; | ||||
| string execute_mode; | |||||
| auto result = ge::GetContext().GetOption(OPTION_EXEC_DYNAMIC_EXECUTE_MODE, execute_mode); | |||||
| if (result != SUCCESS) { | |||||
| GELOGW("Can not get dynamic execute mode attr"); | |||||
| } | |||||
| GELOGD("The dynamic execute is %s", execute_mode.c_str()); | |||||
| string is_copy_output_addr; | |||||
| result = ge::GetContext().GetOption(OPTION_EXEC_ENABLE_COPY_OUTPUT_ADDR, is_copy_output_addr); | |||||
| if (result != SUCCESS) { | |||||
| GELOGW("Can not get option exec enable copy output addr attr"); | |||||
| } | |||||
| GELOGD("The enable copy output addrs is %s", is_copy_output_addr.c_str()); | |||||
| if (execute_mode == kLazyRecompile && is_copy_output_addr == kIsCopyOutputAddr) { | |||||
| GE_IF_BOOL_EXEC(SUCCESS != (ret = CheckAndReleaseStreamEventResource(ge_model, model_id)), | |||||
| GELOGW("[Release][Resource] failed, model id:%u", model_id)); | |||||
| GELOGI("Release stream and event resources for model id:%u success", model_id); | |||||
| } | |||||
| GE_TIMESTAMP_START(Init); | GE_TIMESTAMP_START(Init); | ||||
| GE_IF_BOOL_EXEC(SUCCESS != (ret = davinci_model->Init()), GELOGW("DavinciInit failed."); break;); | GE_IF_BOOL_EXEC(SUCCESS != (ret = davinci_model->Init()), GELOGW("DavinciInit failed."); break;); | ||||
| GE_TIMESTAMP_END(Init, "GraphLoader::ModelInit"); | GE_TIMESTAMP_END(Init, "GraphLoader::ModelInit"); | ||||
| @@ -382,6 +406,126 @@ Status ModelManager::LoadModelOnline(uint32_t &model_id, const shared_ptr<ge::Ge | |||||
| return ret; | return ret; | ||||
| } | } | ||||
| Status ModelManager::CheckAndReleaseStreamEventResource(const GeModelPtr &ge_model, uint32_t model_id) { | |||||
| GE_CHECK_NOTNULL(ge_model); | |||||
| int64_t value = 0; | |||||
| bool ret = ge::AttrUtils::GetInt(ge_model, ATTR_MODEL_STREAM_NUM, value); | |||||
| int64_t need_stream_num = ret? value : 0; | |||||
| ret = ge::AttrUtils::GetInt(ge_model, ATTR_MODEL_EVENT_NUM, value); | |||||
| int64_t need_event_num = ret? value : 0; | |||||
| GELOGD("The main stream number is %lu, event number is %lu", need_stream_num, need_event_num); | |||||
| int64_t hccl_follow_stream = 0; | |||||
| Status status = ModelUtils::CalculateFollowStream(ge_model, hccl_follow_stream); | |||||
| if (status != SUCCESS) { | |||||
| GELOGE(FAILED, "[Calculate][stream] Calculate follow stream num failed"); | |||||
| return FAILED; | |||||
| } | |||||
| need_stream_num = need_stream_num + hccl_follow_stream; | |||||
| GELOGD("The model is %u, need stream is %ld", model_id, need_stream_num); | |||||
| int64_t free_stream_num = 0; | |||||
| status = GetFreeStream(free_stream_num); | |||||
| if (status != SUCCESS) { | |||||
| GELOGE(FAILED, "Get free stream num failed"); | |||||
| return FAILED; | |||||
| } | |||||
| if (need_stream_num > free_stream_num) { | |||||
| status = ReleaseResource(need_stream_num, free_stream_num, kStreamResource); | |||||
| if (status != SUCCESS) { | |||||
| GELOGE(FAILED, "Release stream resoure failed"); | |||||
| return FAILED; | |||||
| } | |||||
| } | |||||
| int64_t free_event_num = 0; | |||||
| GetFreeEvent(free_event_num); | |||||
| if (need_event_num > free_event_num) { | |||||
| status = ReleaseResource(need_event_num, free_event_num, kEventResource); | |||||
| if (status != SUCCESS) { | |||||
| GELOGE(FAILED, "Release event resource failed"); | |||||
| return FAILED; | |||||
| } | |||||
| } | |||||
| return SUCCESS; | |||||
| } | |||||
| Status ModelManager::ReleaseResource(int64_t need_resource, int64_t free_resource, const string &resource_kind) { | |||||
| while (need_resource > free_resource) { | |||||
| uint32_t max_stream_model_id = 0; | |||||
| uint32_t max_event_model_id = 0; | |||||
| GetMaxStreamAndEventModel(max_stream_model_id, max_event_model_id); | |||||
| GELOGD("The max stream num model is: %u, the max event num model is :%u", max_stream_model_id, max_event_model_id); | |||||
| std::lock_guard<std::recursive_mutex> lock(map_mutex_); | |||||
| if (resource_kind == kStreamResource) { | |||||
| uint64_t max_stream_num = model_map_.at(max_stream_model_id)->GetAllStreamNum(); | |||||
| Status ret = Unload(max_stream_model_id); | |||||
| if (ret != SUCCESS) { | |||||
| GELOGE(FAILED, "Unload max stream model failed, model id : %u",max_stream_model_id); | |||||
| return FAILED; | |||||
| } | |||||
| free_resource = free_resource + max_stream_num; | |||||
| GELOGD("Unload model for stream, model id : %u, stream num : %lu", max_stream_model_id, max_stream_num); | |||||
| } | |||||
| if (resource_kind == kEventResource) { | |||||
| uint64_t max_event_num = model_map_.at(max_event_model_id)->GetEventList().size(); | |||||
| Status ret = Unload(max_event_model_id); | |||||
| if (ret != SUCCESS) { | |||||
| GELOGE(FAILED, "Unload max event model failed, model id : %u", max_event_model_id); | |||||
| return FAILED; | |||||
| } | |||||
| free_resource = free_resource + max_event_num; | |||||
| GELOGD("Unload model for event, model id : %u, event num : %zu", max_event_model_id, max_event_num); | |||||
| } | |||||
| } | |||||
| return SUCCESS; | |||||
| } | |||||
| Status ModelManager::GetFreeStream(int64_t &free_stream) { | |||||
| uint32_t max_stream_cout; | |||||
| uint32_t max_task_cout; | |||||
| rtError_t ret = rtGetMaxStreamAndTask(RT_NORMAL_STREAM, &max_stream_cout, &max_task_cout); | |||||
| if (ret != RT_ERROR_NONE) { | |||||
| REPORT_INNER_ERROR("E19999", "Call rtGetMaxStreamAndTask failed"); | |||||
| GELOGE(FAILED, "Get max stream and task cout failed"); | |||||
| return FAILED; | |||||
| } | |||||
| GELOGD("Allowed max stream cout :%u, maxi task cout per stream:%u", max_stream_cout, max_task_cout); | |||||
| std::lock_guard<std::recursive_mutex> lock(map_mutex_); | |||||
| uint64_t stream_sum = 0; | |||||
| for (auto &it : model_map_) { | |||||
| stream_sum = stream_sum + it.second->GetAllStreamNum(); | |||||
| } | |||||
| free_stream = max_stream_cout - stream_sum; | |||||
| return SUCCESS; | |||||
| } | |||||
| void ModelManager::GetFreeEvent(int64_t &free_event) { | |||||
| std::lock_guard<std::recursive_mutex> lock(map_mutex_); | |||||
| uint64_t event_sum = 0; | |||||
| for (auto &it : model_map_) { | |||||
| event_sum = event_sum + it.second->GetEventList().size(); | |||||
| } | |||||
| free_event = kMaxEventNum - event_sum; | |||||
| } | |||||
| void ModelManager::GetMaxStreamAndEventModel(uint32_t &max_stream_model, uint32_t &max_event_model) { | |||||
| std::lock_guard<std::recursive_mutex> lock(map_mutex_); | |||||
| uint64_t max_stream_num = 0; | |||||
| uint64_t max_event_num = 0; | |||||
| for (auto &it : model_map_) { | |||||
| if (it.second ->GetAllStreamNum() > max_stream_num) { | |||||
| max_stream_num = it.second->GetAllStreamNum(); | |||||
| max_stream_model = it.first; | |||||
| } | |||||
| if (it.second->GetEventList().size() > max_event_num) { | |||||
| max_event_num = it.second->GetEventList().size(); | |||||
| max_event_model = it.first; | |||||
| } | |||||
| } | |||||
| } | |||||
| void ModelManager::InsertModel(uint32_t model_id, std::shared_ptr<DavinciModel> &davinci_model) { | void ModelManager::InsertModel(uint32_t model_id, std::shared_ptr<DavinciModel> &davinci_model) { | ||||
| GE_CHK_BOOL_EXEC(davinci_model != nullptr, return, "[Check][Param] davinci_model ptr is null, id:%u", model_id); | GE_CHK_BOOL_EXEC(davinci_model != nullptr, return, "[Check][Param] davinci_model ptr is null, id:%u", model_id); | ||||
| std::lock_guard<std::recursive_mutex> lock(map_mutex_); | std::lock_guard<std::recursive_mutex> lock(map_mutex_); | ||||
| @@ -337,6 +337,11 @@ class FMK_FUNC_HOST_VISIBILITY FMK_FUNC_DEV_VISIBILITY ModelManager { | |||||
| void InsertModel(uint32_t model_id, std::shared_ptr<DavinciModel> &davinci_model); | void InsertModel(uint32_t model_id, std::shared_ptr<DavinciModel> &davinci_model); | ||||
| void InsertModel(uint32_t model_id, std::shared_ptr<hybrid::HybridDavinciModel> &hybrid_model); | void InsertModel(uint32_t model_id, std::shared_ptr<hybrid::HybridDavinciModel> &hybrid_model); | ||||
| Status CheckAndReleaseStreamEventResource(const GeModelPtr &ge_model, uint32_t model_id); | |||||
| Status ReleaseResource(int64_t need_resource, int64_t free_resource, const string &resource_kind); | |||||
| Status GetFreeStream(int64_t &free_stream); | |||||
| void GetFreeEvent(int64_t &free_event); | |||||
| void GetMaxStreamAndEventModel(uint32_t &max_stream_model, uint32_t &max_event_model); | |||||
| /// | /// | ||||
| /// @ingroup domi_ome | /// @ingroup domi_ome | ||||
| /// @brief delete model from model manager set | /// @brief delete model from model manager set | ||||
| @@ -43,6 +43,10 @@ | |||||
| } \ | } \ | ||||
| } while (0) | } while (0) | ||||
| namespace { | |||||
| const char *const kUsedStreamNum = "used_stream_num"; | |||||
| } // namespace | |||||
| namespace ge { | namespace ge { | ||||
| /// | /// | ||||
| /// @ingroup ge | /// @ingroup ge | ||||
| @@ -620,4 +624,54 @@ Status ModelUtils::GetRtAddress(const RuntimeParam ¶m, uintptr_t logic_addr, | |||||
| mem_addr = runtime_base_addr + logic_addr; | mem_addr = runtime_base_addr + logic_addr; | ||||
| return SUCCESS; | return SUCCESS; | ||||
| } | } | ||||
| Status ModelUtils::CalculateFollowStream(const GeModelPtr &ge_model, int64_t &hccl_fellow_stream_num) { | |||||
| const auto &model_def = ge_model->GetModelTaskDefPtr(); | |||||
| GE_CHECK_NOTNULL(model_def); | |||||
| Graph graph = ge_model->GetGraph(); | |||||
| ComputeGraphPtr compute_graph = GraphUtils::GetComputeGraph(graph); | |||||
| GE_CHECK_NOTNULL(compute_graph); | |||||
| map<uint32_t, OpDescPtr> op_list; | |||||
| for (const auto &node : compute_graph->GetDirectNode()) { | |||||
| OpDescPtr op_desc = node->GetOpDesc(); | |||||
| GE_CHECK_NOTNULL(op_desc); | |||||
| op_list.emplace(op_desc->GetId(), op_desc); | |||||
| } | |||||
| std::multimap<int64_t, int64_t> main_follow_num; | |||||
| for (int32_t i = 0; i < model_def->task_size(); i++) { | |||||
| const domi::TaskDef &task = model_def->task(i); | |||||
| if (static_cast<rtModelTaskType_t>(task.type() == RT_MODEL_TASK_HCCL)){ | |||||
| auto hccl_def = task.kernel_hccl(); | |||||
| OpDescPtr hccl_op_desc = op_list.at(hccl_def.op_index()); | |||||
| int64_t main_stream_id = hccl_op_desc->GetStreamId(); | |||||
| int64_t follow_stream_num = 0; | |||||
| if (!ge::AttrUtils::GetInt(hccl_op_desc, kUsedStreamNum, follow_stream_num)) { | |||||
| GELOGW("Get used stream num failed, op is %s", hccl_op_desc->GetName().c_str()); | |||||
| } | |||||
| main_follow_num.emplace(main_stream_id, follow_stream_num); | |||||
| } | |||||
| } | |||||
| hccl_fellow_stream_num = CalFollowStreamSum(main_follow_num); | |||||
| return SUCCESS; | |||||
| } | |||||
| int64_t ModelUtils::CalFollowStreamSum(const std::multimap<int64_t, int64_t> &hccl_stream_map) { | |||||
| std::map<int64_t, int64_t> max_follow_stream_map; | |||||
| for (const auto &it : hccl_stream_map) { | |||||
| auto max_it = max_follow_stream_map.find(it.first); | |||||
| if(max_it == max_follow_stream_map.end()) { | |||||
| max_follow_stream_map.emplace(it.first, it.second); | |||||
| } else if (it.second > max_it->second) { | |||||
| max_follow_stream_map.at(max_it->first) = it.second; | |||||
| } | |||||
| } | |||||
| int64_t need_follow_stream_num = 0; | |||||
| for (const auto &follow_it : max_follow_stream_map) { | |||||
| need_follow_stream_num = need_follow_stream_num + follow_it.second; | |||||
| } | |||||
| GELOGD("Need follow num is %ld", need_follow_stream_num); | |||||
| return need_follow_stream_num; | |||||
| } | |||||
| } // namespace ge | } // namespace ge | ||||
| @@ -24,6 +24,7 @@ | |||||
| #include "graph/load/model_manager/task_info/task_info.h" | #include "graph/load/model_manager/task_info/task_info.h" | ||||
| #include "graph/op_desc.h" | #include "graph/op_desc.h" | ||||
| #include "graph/utils/tensor_adapter.h" | #include "graph/utils/tensor_adapter.h" | ||||
| #include "common/model/ge_model.h" | |||||
| using std::vector; | using std::vector; | ||||
| @@ -108,6 +109,19 @@ class ModelUtils { | |||||
| /// | /// | ||||
| static Status GetRtAddress(const RuntimeParam &model_param, uintptr_t logic_addr, uint8_t *&mem_addr); | static Status GetRtAddress(const RuntimeParam &model_param, uintptr_t logic_addr, uint8_t *&mem_addr); | ||||
| /// | |||||
| /// @ingroup ge | |||||
| /// @brief Calculate hccl follw stream | |||||
| /// @return Status | |||||
| /// | |||||
| static Status CalculateFollowStream(const GeModelPtr &ge_model, int64_t &hccl_fellow_stream_num); | |||||
| /// | |||||
| /// @ingroup ge | |||||
| /// @brief Calculate the sum of follow stream | |||||
| /// @return int64_t | |||||
| /// | |||||
| static int64_t CalFollowStreamSum(const std::multimap<int64_t, int64_t> &hccl_stream_map); | |||||
| private: | private: | ||||
| /// | /// | ||||
| /// @ingroup ge | /// @ingroup ge | ||||
| @@ -23,7 +23,7 @@ | |||||
| #include "external/ge/ge_api_types.h" | #include "external/ge/ge_api_types.h" | ||||
| #include "common/dump/opdebug_register.h" | #include "common/dump/opdebug_register.h" | ||||
| #include "graph/load/model_manager/data_inputer.h" | #include "graph/load/model_manager/data_inputer.h" | ||||
| #include "graph/load/model_manager/data_dumper.h" | |||||
| #include "common/dump/data_dumper.h" | |||||
| #include "hybrid/executor/hybrid_model_executor.h" | #include "hybrid/executor/hybrid_model_executor.h" | ||||
| #include "hybrid/executor/hybrid_model_pipeline_executor.h" | #include "hybrid/executor/hybrid_model_pipeline_executor.h" | ||||
| #include "runtime/stream.h" | #include "runtime/stream.h" | ||||
| @@ -67,6 +67,7 @@ const char *const OPTION_EXEC_ENABLE_TAILING_OPTIMIZATION = "ge.exec.isTailingOp | |||||
| const char *const OPTION_EXEC_DYNAMIC_INPUT = "ge.exec.dynamicInput"; | const char *const OPTION_EXEC_DYNAMIC_INPUT = "ge.exec.dynamicInput"; | ||||
| const char *const OPTION_EXEC_DYNAMIC_EXECUTE_MODE = "ge.exec.dynamicGraphExecuteMode"; | const char *const OPTION_EXEC_DYNAMIC_EXECUTE_MODE = "ge.exec.dynamicGraphExecuteMode"; | ||||
| const char *const OPTION_EXEC_DATA_INPUTS_SHAPE_RANGE = "ge.exec.dataInputsShapeRange"; | const char *const OPTION_EXEC_DATA_INPUTS_SHAPE_RANGE = "ge.exec.dataInputsShapeRange"; | ||||
| const char *const OPTION_EXEC_ENABLE_COPY_OUTPUT_ADDR = "ge.exec.enableCopyOutputAddr"; | |||||
| // Option key: memory init | // Option key: memory init | ||||
| const char *const GRAPH_MEMORY_MAX_SIZE = "ge.graphMemoryMaxSize"; | const char *const GRAPH_MEMORY_MAX_SIZE = "ge.graphMemoryMaxSize"; | ||||
| @@ -441,8 +442,8 @@ const std::set<std::string> ir_builder_suppported_options = {INPUT_FORMAT, | |||||
| // for interface: aclgrphParse | // for interface: aclgrphParse | ||||
| const std::set<std::string> ir_parser_suppported_options = { | const std::set<std::string> ir_parser_suppported_options = { | ||||
| INPUT_FP16_NODES, IS_INPUT_ADJUST_HW_LAYOUT, IS_OUTPUT_ADJUST_HW_LAYOUT, OUTPUT, | |||||
| OUT_NODES, ENABLE_SCOPE_FUSION_PASSES}; | |||||
| INPUT_FP16_NODES, IS_INPUT_ADJUST_HW_LAYOUT, IS_OUTPUT_ADJUST_HW_LAYOUT, OUTPUT, | |||||
| OUT_NODES, ENABLE_SCOPE_FUSION_PASSES}; | |||||
| // for interface: aclgrphBuildInitialize | // for interface: aclgrphBuildInitialize | ||||
| const std::set<std::string> global_options = {CORE_TYPE, | const std::set<std::string> global_options = {CORE_TYPE, | ||||
| @@ -218,7 +218,7 @@ set(GRAPH_PREPARE_COMMON_SRC_FILES | |||||
| set(GRAPH_DAVINCI_MODEL_SRC_FILES | set(GRAPH_DAVINCI_MODEL_SRC_FILES | ||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/aipp_utils.cc" | "${GE_CODE_DIR}/ge/graph/load/model_manager/aipp_utils.cc" | ||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/cpu_queue_schedule.cc" | "${GE_CODE_DIR}/ge/graph/load/model_manager/cpu_queue_schedule.cc" | ||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/data_dumper.cc" | |||||
| "${GE_CODE_DIR}/ge/common/dump/data_dumper.cc" | |||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/data_inputer.cc" | "${GE_CODE_DIR}/ge/graph/load/model_manager/data_inputer.cc" | ||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/davinci_model.cc" | "${GE_CODE_DIR}/ge/graph/load/model_manager/davinci_model.cc" | ||||
| "${GE_CODE_DIR}/ge/graph/load/model_manager/model_manager.cc" | "${GE_CODE_DIR}/ge/graph/load/model_manager/model_manager.cc" | ||||
| @@ -18,7 +18,7 @@ | |||||
| #define private public | #define private public | ||||
| #define protected public | #define protected public | ||||
| #include "graph/load/model_manager/data_dumper.h" | |||||
| #include "common/dump/data_dumper.h" | |||||
| #include "graph/load/model_manager/davinci_model.h" | #include "graph/load/model_manager/davinci_model.h" | ||||
| #undef private | #undef private | ||||
| #undef protected | #undef protected | ||||
| @@ -57,6 +57,23 @@ class UtestModelManagerModelManager : public testing::Test { | |||||
| void SetUp() {} | void SetUp() {} | ||||
| void TearDown() {} | void TearDown() {} | ||||
| void CreateGraph(Graph &graph) { | |||||
| TensorDesc desc(ge::Shape({1, 3, 224, 224})); | |||||
| uint32_t size = desc.GetShape().GetShapeSize(); | |||||
| desc.SetSize(size); | |||||
| auto data = op::Data("Data").set_attr_index(0); | |||||
| data.update_input_desc_data(desc); | |||||
| data.update_output_desc_out(desc); | |||||
| auto flatten = op::Flatten("Flatten").set_input_x(data, data.name_out_out()); | |||||
| std::vector<Operator> inputs{data}; | |||||
| std::vector<Operator> outputs{flatten}; | |||||
| std::vector<Operator> targets{flatten}; | |||||
| // Graph graph("test_graph"); | |||||
| graph.SetInputs(inputs).SetOutputs(outputs).SetTargets(targets); | |||||
| } | |||||
| void GenUnencryptModelData(ModelData &data) { | void GenUnencryptModelData(ModelData &data) { | ||||
| const int model_len = 10; | const int model_len = 10; | ||||
| data.model_len = sizeof(ModelFileHeader) + model_len; | data.model_len = sizeof(ModelFileHeader) + model_len; | ||||
| @@ -454,4 +471,74 @@ TEST_F(UtestModelManagerModelManager, command_profiling) { | |||||
| Status ret = manager.HandleProfModelUnsubscribeCommand(cmd); | Status ret = manager.HandleProfModelUnsubscribeCommand(cmd); | ||||
| profiling_manager.CleanSubscribeInfo(); | profiling_manager.CleanSubscribeInfo(); | ||||
| } | } | ||||
| TEST_F(UtestModelManagerModelManager, Cal_follow_stream_sum) { | |||||
| std::multimap<int64_t, int64_t> hccl_stream_map = {{1,10}, {1,20}, {2,10}, {2,5}}; | |||||
| int64_t result = ModelUtils::CalFollowStreamSum(hccl_stream_map); | |||||
| EXPECT_EQ(result, 30); | |||||
| } | |||||
| TEST_F(UtestModelManagerModelManager, get_max_stream_and_event) { | |||||
| ModelManager mm; | |||||
| auto model1 = std::make_shared<DavinciModel> (1, nullptr); | |||||
| auto model2 = std::make_shared<DavinciModel> (2, nullptr); | |||||
| rtStream_t stream = nullptr; | |||||
| rtStream_t stream2 = nullptr; | |||||
| rtStream_t stream3 = nullptr; | |||||
| rtStream_t stream4 = nullptr; | |||||
| rtEvent_t event = nullptr; | |||||
| rtEvent_t event2 = nullptr; | |||||
| rtEvent_t event3 = nullptr; | |||||
| model1->stream_list_ = {stream, stream2, stream3, stream4}; | |||||
| model1->event_list_ = {event, event2}; | |||||
| model2->stream_list_ = {stream, stream2}; | |||||
| model2->event_list_ = {event, event2, event3}; | |||||
| mm.InsertModel(1, model1); | |||||
| mm.InsertModel(2, model2); | |||||
| uint32_t max_stream_model; | |||||
| uint32_t max_event_model; | |||||
| mm.GetMaxStreamAndEventModel(max_stream_model, max_event_model); | |||||
| EXPECT_EQ(max_stream_model, 1); | |||||
| EXPECT_EQ(max_event_model, 2); | |||||
| int64_t free_stream; | |||||
| int64_t free_event; | |||||
| Status ret = mm.GetFreeStream(free_stream); | |||||
| EXPECT_EQ(ret, SUCCESS); | |||||
| } | |||||
| TEST_F(UtestModelManagerModelManager, release_resource_stream) { | |||||
| ModelManager mm; | |||||
| auto model1 = std::make_shared<DavinciModel> (1, nullptr); | |||||
| auto model2 = std::make_shared<DavinciModel> (2, nullptr); | |||||
| rtStream_t stream = nullptr; | |||||
| rtStream_t stream2 = nullptr; | |||||
| rtStream_t stream3 = nullptr; | |||||
| rtStream_t stream4 = nullptr; | |||||
| rtEvent_t event = nullptr; | |||||
| rtEvent_t event2 = nullptr; | |||||
| rtEvent_t event3 = nullptr; | |||||
| model1->stream_list_ = {stream, stream2, stream3, stream4}; | |||||
| model1->event_list_ = {event, event2}; | |||||
| model2->stream_list_ = {stream, stream2}; | |||||
| model2->event_list_ = {event, event2, event3}; | |||||
| mm.InsertModel(1, model1); | |||||
| mm.InsertModel(2, model2); | |||||
| string kind = "stream"; | |||||
| Status ret = mm.ReleaseResource(110, 109, kind); | |||||
| EXPECT_EQ(ret, SUCCESS); | |||||
| string kind2 = "event"; | |||||
| Status ret2 = mm.ReleaseResource(110, 109, kind2); | |||||
| EXPECT_EQ(ret2, SUCCESS); | |||||
| } | |||||
| TEST_F(UtestModelManagerModelManager, check_stream_and_event_resource) { | |||||
| ModelManager mm; | |||||
| auto ge_model = make_shared<GeModel>(); | |||||
| Status ret = mm.CheckAndReleaseStreamEventResource(ge_model, 1); | |||||
| EXPECT_EQ(ret, FAILED); | |||||
| } | |||||
| } // namespace ge | } // namespace ge | ||||