From: @shenwei41 Reviewed-by: @liujunzhu,@lilongfei15 Signed-off-by: @lilongfei15tags/v1.2.0
| @@ -258,6 +258,7 @@ set(TRAIN_SRC_LIST | |||
| "graph/passes/get_original_format_pass.cc" | |||
| "graph/passes/guarantee_const_pass.cc" | |||
| "graph/passes/hccl_memcpy_pass.cc" | |||
| "graph/passes/hccl_continuous_memcpy_pass.cc" | |||
| "graph/passes/identity_pass.cc" | |||
| "graph/passes/ref_identity_delete_op_pass.cc" | |||
| "graph/passes/infershape_pass.cc" | |||
| @@ -595,6 +596,7 @@ set(INFER_SRC_LIST | |||
| "graph/passes/cast_remove_pass.cc" | |||
| "graph/passes/transpose_transdata_pass.cc" | |||
| "graph/passes/hccl_memcpy_pass.cc" | |||
| "graph/passes/hccl_continuous_memcpy_pass.cc" | |||
| "graph/passes/flow_ctrl_pass.cc" | |||
| "graph/passes/global_step_insert_pass.cc" | |||
| "graph/passes/link_gen_mask_nodes_pass.cc" | |||
| @@ -707,7 +709,7 @@ target_compile_options(ge_runner PRIVATE | |||
| -O2 | |||
| -fno-common | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-variable> | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-const-variable> | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-const-variable -Werror=format> | |||
| ) | |||
| target_include_directories(ge_runner SYSTEM PRIVATE | |||
| @@ -776,7 +778,7 @@ target_compile_options(ge_compiler PRIVATE | |||
| -O2 | |||
| -fno-common | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-variable> | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-const-variable> | |||
| $<$<STREQUAL:${CMAKE_CXX_COMPILER_VERSION},7.3.0>:-Werror=unused-const-variable -Werror=format> | |||
| ) | |||
| target_include_directories(ge_compiler SYSTEM PRIVATE | |||
| @@ -212,6 +212,7 @@ OMG_HOST_SRC_FILES := \ | |||
| graph/passes/cast_remove_pass.cc \ | |||
| graph/passes/transpose_transdata_pass.cc \ | |||
| graph/passes/hccl_memcpy_pass.cc \ | |||
| graph/passes/hccl_continuous_memcpy_pass.cc \ | |||
| graph/passes/flow_ctrl_pass.cc \ | |||
| graph/passes/global_step_insert_pass.cc \ | |||
| graph/passes/link_gen_mask_nodes_pass.cc \ | |||
| @@ -183,6 +183,7 @@ LIBGE_LOCAL_SRC_FILES := \ | |||
| graph/passes/get_original_format_pass.cc \ | |||
| graph/passes/guarantee_const_pass.cc \ | |||
| graph/passes/hccl_memcpy_pass.cc \ | |||
| graph/passes/hccl_continuous_memcpy_pass.cc \ | |||
| graph/passes/identity_pass.cc \ | |||
| graph/passes/ref_identity_delete_op_pass.cc \ | |||
| graph/passes/infershape_pass.cc \ | |||
| @@ -47,6 +47,7 @@ const char *const kEngineNameDefault = "default"; | |||
| const char *const kVectorEngine = "VectorEngine"; | |||
| const char *const kAIcoreEngine = "AIcoreEngine"; | |||
| const char *const kFileNameSuffix = "online"; | |||
| const char *const kAicpuAllshape = "_AllShape"; | |||
| const size_t kDynamicDimSize = 1; | |||
| const int64_t kDynamicDimValue = -2; | |||
| @@ -721,8 +722,12 @@ Status GeGenerator::BuildSingleOp(OpDescPtr &op_desc, const vector<GeTensor> &in | |||
| GeModelPtr &ge_model = name_to_ge_model.begin()->second; | |||
| GELOGD("The opType in op_desc_tmp is [%s]", op_desc_tmp->GetType().c_str()); | |||
| bool all_shape = false; | |||
| bool dynamic_flag = false; | |||
| if (CheckShapeReset(op_desc, dynamic_flag) == SUCCESS && dynamic_flag) { | |||
| (void)AttrUtils::GetBool(op_desc, kAicpuAllshape, all_shape); | |||
| CheckShapeReset(op_desc, dynamic_flag); | |||
| if (dynamic_flag || all_shape) { | |||
| GELOGD("Get aicpu all_shape kernel!"); | |||
| vector<GeTensor> inputs_dynamic; | |||
| vector<GeTensor> outputs_dynamic; | |||
| GE_CHK_STATUS_RET_NOLOG(ResetTensorVecShape(inputs, inputs_dynamic)); | |||
| @@ -88,6 +88,14 @@ Status VariableMemoryAssigner::AssignVarAttr2Nodes() { | |||
| return ge::SUCCESS; | |||
| } | |||
| Status VariableMemoryAssigner::AssignMemory2HasRefAttrNode() { | |||
| Status result = ge::VarMemAssignUtil::AssignMemory2HasRefAttrNode(compute_graph_); | |||
| if (result != ge::SUCCESS) { | |||
| return result; | |||
| } | |||
| return ge::SUCCESS; | |||
| } | |||
| Status GraphMemoryAssigner::AssignMemory() { | |||
| ge::HybridMemAssignerPtr mem_assigner(new(std::nothrow) HybridMemAssigner(compute_graph_)); | |||
| if (mem_assigner->Assign() != ge::SUCCESS) { | |||
| @@ -135,6 +143,19 @@ ge::Status GraphMemoryAssigner::AssignVarAttr2Nodes() { | |||
| return ge::SUCCESS; | |||
| } | |||
| ge::Status GraphMemoryAssigner::AssignMemory2HasRefAttrNode() { | |||
| auto variable_assigner = | |||
| std::unique_ptr<ge::VariableMemoryAssigner>(new(std::nothrow) ge::VariableMemoryAssigner(compute_graph_)); | |||
| if (variable_assigner == nullptr) { | |||
| GELOGE(ge::FAILED, "Alloc VariableMemoryAssigner failed."); | |||
| return ge::FAILED; | |||
| } | |||
| if (variable_assigner->AssignMemory2HasRefAttrNode() != ge::SUCCESS) { | |||
| return ge::FAILED; | |||
| } | |||
| return ge::SUCCESS; | |||
| } | |||
| ge::Status CalculateTensorRealSizeAndOutSize(const ge::ConstGeTensorDescPtr &output_desc, | |||
| int64_t dim_index, int64_t &output_mem_size, | |||
| int64_t &batch_dim_num, int64_t &out_size) { | |||
| @@ -63,6 +63,8 @@ class VariableMemoryAssigner { | |||
| /// | |||
| ge::Status AssignVarAttr2Nodes(); | |||
| ge::Status AssignMemory2HasRefAttrNode(); | |||
| private: | |||
| ge::ComputeGraphPtr compute_graph_; | |||
| }; | |||
| @@ -99,6 +101,8 @@ class GraphMemoryAssigner { | |||
| /// | |||
| ge::Status AssignVarAttr2Nodes(); | |||
| ge::Status AssignMemory2HasRefAttrNode(); | |||
| ge::Status ReAssignMemory(bool is_loop_graph, map<int64_t, size_t> &mem_type_to_offset); | |||
| ge::Status AssignZeroCopyMemory(map<int64_t, size_t> &mem_offset, size_t &zero_mem_copy_size); | |||
| @@ -40,6 +40,11 @@ Status MemoryAssigner::AssignMemory(bool is_loop_graph, map<int64_t, size_t> &me | |||
| return ge::FAILED; | |||
| } | |||
| if (graph_mem_assigner.AssignMemory2HasRefAttrNode() != ge::SUCCESS) { | |||
| GELOGE(ge::FAILED, "Assign memory to node which has ref attr failed!"); | |||
| return ge::FAILED; | |||
| } | |||
| // Assign memory for reference | |||
| if (graph_mem_assigner.AssignReferenceMemory() != ge::SUCCESS) { | |||
| GELOGE(ge::FAILED, "Assign reference memory failed!"); | |||
| @@ -33,10 +33,7 @@ using std::vector; | |||
| namespace ge { | |||
| Status VarMemAssignUtil::AssignVarMemory(ge::ComputeGraphPtr &compute_graph) { | |||
| GE_CHK_STATUS_RET(AssignMemory2VariableNode(compute_graph)); | |||
| GE_CHK_STATUS_RET(AssignMemory2HasRefAttrNode(compute_graph)); | |||
| return SUCCESS; | |||
| return AssignMemory2VariableNode(compute_graph); | |||
| } | |||
| Status VarMemAssignUtil::AssignConstantOpMemory(ge::ComputeGraphPtr &compute_graph) { | |||
| @@ -446,23 +446,20 @@ void DavinciModel::InitRuntimeParams() { | |||
| runtime_param_.mem_size, runtime_param_.weight_size, runtime_param_.var_size); | |||
| } | |||
| void DavinciModel::CheckHasHcomOp() { | |||
| Graph graph = ge_model_->GetGraph(); | |||
| auto compute_graph = GraphUtils::GetComputeGraph(graph); | |||
| if (compute_graph == nullptr) { | |||
| return; | |||
| } | |||
| void DavinciModel::CheckHasHcomOp(const ComputeGraphPtr &compute_graph) { | |||
| const set<string> hcom_opp_types({ | |||
| HCOMBROADCAST, HCOMALLGATHER, HCOMALLREDUCE, HCOMSEND, HCOMRECEIVE, HCOMREDUCESCATTER, | |||
| HVDCALLBACKALLREDUCE, HVDCALLBACKALLGATHER, HVDCALLBACKBROADCAST, HVDWAIT, HCOMREDUCE | |||
| }); | |||
| for (const auto &node : compute_graph->GetAllNodes()) { | |||
| OpDescPtr op_desc = node->GetOpDesc(); | |||
| GE_IF_BOOL_EXEC(op_desc == nullptr, GELOGW("Node OpDesc is nullptr"); continue); | |||
| GE_IF_BOOL_EXEC(((op_desc->GetType() == HCOMBROADCAST) || (op_desc->GetType() == HCOMALLGATHER) || | |||
| (op_desc->GetType() == HCOMALLREDUCE) || (op_desc->GetType() == HCOMSEND) || | |||
| (op_desc->GetType() == HCOMRECEIVE) || (op_desc->GetType() == HCOMREDUCESCATTER) || | |||
| (op_desc->GetType() == HVDCALLBACKALLREDUCE) || (op_desc->GetType() == HVDCALLBACKALLGATHER) || | |||
| (op_desc->GetType() == HVDCALLBACKBROADCAST) || (op_desc->GetType() == HVDWAIT) || | |||
| (op_desc->GetType() == HCOMREDUCE)), | |||
| uint32_t stream_id = static_cast<uint32_t>(op_desc->GetStreamId()); | |||
| (void)hcom_streams_.emplace(stream_id); GELOGD("hcom stream: %u.", stream_id); continue); | |||
| if (hcom_opp_types.count(op_desc->GetType()) > 0) { | |||
| uint32_t stream_id = static_cast<uint32_t>(op_desc->GetStreamId()); | |||
| hcom_streams_.emplace(stream_id); | |||
| GELOGD("hcom stream: %u.", stream_id); | |||
| } | |||
| } | |||
| } | |||
| @@ -642,7 +639,7 @@ Status DavinciModel::Init(void *dev_ptr, size_t mem_size, void *weight_ptr, size | |||
| name_ = ge_model_->GetName(); | |||
| (void)ge::AttrUtils::GetBool(ge_model_, ATTR_NAME_SWITCH_FOR_L1_FUSION, is_l1_fusion_enable_); | |||
| GELOGD("The value of ge.l1Fusion in ge_model is %d.", is_l1_fusion_enable_); | |||
| CheckHasHcomOp(); | |||
| CheckHasHcomOp(compute_graph); | |||
| vector<int64_t> huge_stream_list; | |||
| (void)ge::AttrUtils::GetListInt(ge_model_, ATTR_MODEL_HUGE_STREAM_LIST, huge_stream_list); | |||
| @@ -1028,7 +1025,7 @@ Status DavinciModel::GenInputOutputInfo(const map<uint32_t, OpDescPtr> &data_by_ | |||
| const vector<OpDescPtr> &output_op_list) { | |||
| GELOGD("Data node size: %zu, NetOutput node size: %zu", data_by_index.size(), output_op_list.size()); | |||
| for (auto &item : data_by_index) { | |||
| auto output_addrs = ModelUtils::GetOutputDataAddrs(runtime_param_, item.second); | |||
| const auto output_addrs = ModelUtils::GetOutputDataAddrs(runtime_param_, item.second); | |||
| GELOGD("Data node: %s, output addr size: %zu", item.second->GetName().c_str(), output_addrs.size()); | |||
| input_addrs_list_.emplace_back(output_addrs); | |||
| @@ -1036,14 +1033,18 @@ Status DavinciModel::GenInputOutputInfo(const map<uint32_t, OpDescPtr> &data_by_ | |||
| GE_CHK_STATUS_RET(InitAippType(item.first, item.second, data_by_index), "Init AIPP Type failed"); | |||
| GE_CHK_STATUS_RET(InitOrigInputInfo(item.first, item.second), "Init Orig input failed"); | |||
| GE_CHK_STATUS_RET(InitAippInputOutputDims(item.first, item.second), "Init AIPP dims failed"); | |||
| GE_CHK_STATUS_RET(InitInputDescInfo(item.second), "Init input desc info failed"); | |||
| if (item.second->GetType() == AIPP_DATA_TYPE) { | |||
| GELOGI("This is dynamic aipp model, Node: %s", item.second->GetName().c_str()); | |||
| is_dynamic_aipp_ = true; | |||
| } | |||
| } | |||
| vector<string> out_node_name; | |||
| (void)AttrUtils::GetListStr(ge_model_, ATTR_MODEL_OUT_NODES_NAME, out_node_name); | |||
| GELOGD("Output node size: %zu, out nodes name: %zu", output_op_list.size(), out_node_name.size()); | |||
| for (const auto &op_desc : output_op_list) { | |||
| auto input_addrs = ModelUtils::GetInputDataAddrs(runtime_param_, op_desc); | |||
| const auto input_addrs = ModelUtils::GetInputDataAddrs(runtime_param_, op_desc); | |||
| GELOGD("NetOutput node: %s, input addr size: %zu", op_desc->GetName().c_str(), input_addrs.size()); | |||
| output_addrs_list_.emplace_back(input_addrs); | |||
| @@ -1061,10 +1062,11 @@ Status DavinciModel::GenInputOutputInfo(const map<uint32_t, OpDescPtr> &data_by_ | |||
| if (InitOutputTensorInfo(op_desc) != SUCCESS) { | |||
| return INTERNAL_ERROR; | |||
| } | |||
| GE_CHK_STATUS_RET(InitOutputDescInfo(op_desc, out_node_name), "Init output desc info failed"); | |||
| } | |||
| GE_CHK_STATUS_RET(InitInputDescInfo(data_by_index), "Init input desc info failed"); | |||
| return InitOutputDescInfo(output_op_list); | |||
| return SUCCESS; | |||
| } | |||
| bool DavinciModel::IsGetNextSinkDynamic(const OpDescPtr &op_desc) { | |||
| @@ -1980,27 +1982,24 @@ void DavinciModel::CreateInputDimsInfo(const OpDescPtr &op_desc, Format format, | |||
| } | |||
| } | |||
| Status DavinciModel::InitInputDescInfo(const map<uint32_t, OpDescPtr> &data_by_index) { | |||
| for (const auto &item : data_by_index) { | |||
| const auto op_desc = item.second; | |||
| GE_CHECK_NOTNULL(op_desc->GetInputDescPtr(0)); | |||
| Status DavinciModel::InitInputDescInfo(const OpDescPtr &op_desc) { | |||
| GE_CHECK_NOTNULL(op_desc->GetInputDescPtr(0)); | |||
| InputOutputDescInfo input; | |||
| ShapeDescription dims_info; | |||
| Format format = op_desc->GetInputDescPtr(0)->GetFormat(); | |||
| CreateInputDimsInfo(op_desc, format, input.shape_info, dims_info); | |||
| InputOutputDescInfo input; | |||
| ShapeDescription dims_info; | |||
| Format format = op_desc->GetInputDescPtr(0)->GetFormat(); | |||
| CreateInputDimsInfo(op_desc, format, input.shape_info, dims_info); | |||
| input.data_type = op_desc->GetInputDescPtr(0)->GetDataType(); | |||
| input.name = op_desc->GetName(); | |||
| int64_t input_size = 0; | |||
| GE_CHK_STATUS_RET(TensorUtils::GetSize(*op_desc->GetInputDescPtr(0), input_size), "get input size failed."); | |||
| input.size = input_size; | |||
| input_formats_.push_back(format); | |||
| input_descs_.push_back(input); | |||
| input.data_type = op_desc->GetInputDescPtr(0)->GetDataType(); | |||
| input.name = op_desc->GetName(); | |||
| int64_t input_size = 0; | |||
| GE_CHK_STATUS_RET(TensorUtils::GetSize(*op_desc->GetInputDescPtr(0), input_size), "get input size failed."); | |||
| input.size = input_size; | |||
| input_formats_.push_back(format); | |||
| input_descs_.push_back(input); | |||
| input.shape_info = dims_info; | |||
| input_descs_dims_.push_back(input); | |||
| } | |||
| input.shape_info = dims_info; | |||
| input_descs_dims_.push_back(input); | |||
| return SUCCESS; | |||
| } | |||
| @@ -2066,37 +2065,31 @@ void DavinciModel::CreateOutput(uint32_t index, const OpDescPtr &op_desc, InputO | |||
| output.data_type = op_desc->GetInputDescPtr(index)->GetDataType(); | |||
| } | |||
| Status DavinciModel::InitOutputDescInfo(const vector<OpDescPtr> &output_op_list) { | |||
| GELOGD("Output node size: %zu", output_op_list.size()); | |||
| vector<string> out_node_name; | |||
| (void)ge::AttrUtils::GetListStr(ge_model_, ATTR_MODEL_OUT_NODES_NAME, out_node_name); | |||
| for (const auto &op_desc : output_op_list) { | |||
| uint32_t out_size = static_cast<uint32_t>(op_desc->GetInputsSize()); | |||
| for (uint32_t index = 0; index < out_size; index++) { | |||
| string output_name; | |||
| InputOutputDescInfo output; | |||
| uint32_t format_result; | |||
| CreateOutput(index, op_desc, output, format_result); | |||
| std::vector<std::string> src_name = op_desc->GetSrcName(); | |||
| std::vector<int64_t> src_index = op_desc->GetSrcIndex(); | |||
| GE_CHK_BOOL_RET_STATUS(src_name.size() > index && src_index.size() > index, INTERNAL_ERROR, | |||
| "construct output_name failed."); | |||
| // forward compatbility, if old om has no out_node_name, need to return output follow origin way | |||
| if (out_size == out_node_name.size()) { | |||
| // neweast plan, the index will add to name during generate model. | |||
| bool contains_colon = out_node_name[index].find(":") != std::string::npos; | |||
| output_name = | |||
| contains_colon ? out_node_name[index] : out_node_name[index] + ":" + std::to_string(src_index[index]); | |||
| } else { | |||
| output_name = std::string("output_") + std::to_string(index) + "_" + src_name[index] + "_" + | |||
| std::to_string(src_index[index]); | |||
| } | |||
| output.name = output_name; | |||
| output_descs_.push_back(output); | |||
| output_formats_.push_back(format_result); | |||
| Status DavinciModel::InitOutputDescInfo(const OpDescPtr &op_desc, const vector<string> &out_node_name) { | |||
| uint32_t out_size = static_cast<uint32_t>(op_desc->GetInputsSize()); | |||
| for (uint32_t i = 0; i < out_size; ++i) { | |||
| string output_name; | |||
| InputOutputDescInfo output; | |||
| uint32_t format_result; | |||
| CreateOutput(i, op_desc, output, format_result); | |||
| std::vector<std::string> src_name = op_desc->GetSrcName(); | |||
| std::vector<int64_t> src_index = op_desc->GetSrcIndex(); | |||
| GE_CHK_BOOL_RET_STATUS(src_name.size() > i && src_index.size() > i, INTERNAL_ERROR, | |||
| "construct output_name failed."); | |||
| // forward compatbility, if old om has no out_node_name, need to return output follow origin way | |||
| if (out_size == out_node_name.size()) { | |||
| // neweast plan, the index will add to name during generate model. | |||
| bool contains_colon = out_node_name[i].find(":") != std::string::npos; | |||
| output_name = contains_colon ? out_node_name[i] : out_node_name[i] + ":" + std::to_string(src_index[i]); | |||
| } else { | |||
| output_name = string("output_") + std::to_string(i) + "_" + src_name[i] + "_" + std::to_string(src_index[i]); | |||
| } | |||
| output.name = output_name; | |||
| output_descs_.push_back(output); | |||
| output_formats_.push_back(format_result); | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| @@ -831,7 +831,7 @@ class DavinciModel { | |||
| void OpDebugUnRegister(); | |||
| void CheckHasHcomOp(); | |||
| void CheckHasHcomOp(const ComputeGraphPtr &graph); | |||
| Status DoTaskSink(); | |||
| @@ -854,8 +854,8 @@ class DavinciModel { | |||
| Status InitOutputTensorInfo(const OpDescPtr &op_desc); | |||
| Status GenOutputTensorInfo(OutputData *output_data, vector<OutputTensorInfo> &outputs); | |||
| Status InitInputDescInfo(const map<uint32_t, OpDescPtr> &data_by_index); | |||
| Status InitOutputDescInfo(const vector<OpDescPtr> &output_op_list); | |||
| Status InitInputDescInfo(const OpDescPtr &op_desc); | |||
| Status InitOutputDescInfo(const OpDescPtr &op_desc, const vector<string> &out_node_name); | |||
| Status InitOrigInputInfo(uint32_t index, const OpDescPtr &op_desc); | |||
| Status InitAippInfo(uint32_t index, const OpDescPtr &op_desc); | |||
| @@ -26,8 +26,42 @@ | |||
| #include "graph/attr_value.h" | |||
| #include "graph/load/model_manager/davinci_model.h" | |||
| #include "graph/load/model_manager/model_manager.h" | |||
| #include "hybrid/node_executor/aicpu/aicpu_ext_info.h" | |||
| #include "framework/common/debug/log.h" | |||
| namespace ge { | |||
| Status KernelExTaskInfo::InitTaskExtInfo(const std::string &ext_info, const OpDescPtr &op_desc) { | |||
| if (ext_info.empty()) { | |||
| return SUCCESS; | |||
| } | |||
| int32_t unknown_shape_type_val = 0; | |||
| (void) AttrUtils::GetInt(op_desc, ::ge::ATTR_NAME_UNKNOWN_SHAPE_TYPE, unknown_shape_type_val); | |||
| UnknowShapeOpType unknown_type = static_cast<UnknowShapeOpType>(unknown_shape_type_val); | |||
| uint32_t num_inputs = op_desc->GetInputsSize(); | |||
| uint32_t num_outputs = op_desc->GetOutputsSize(); | |||
| std::unique_ptr<ge::hybrid::AicpuExtInfoHandler> ext_handle( | |||
| new(std::nothrow) ::ge::hybrid::AicpuExtInfoHandler(op_desc->GetName(), | |||
| num_inputs, | |||
| num_outputs, | |||
| unknown_type)); | |||
| GE_CHK_BOOL_RET_STATUS(ext_handle != nullptr, FAILED, "Malloc aicpu_ext_handle mem failed!"); | |||
| GE_CHK_STATUS_RET(ext_handle->Parse(ext_info), | |||
| "Parse kernel ext info failed, kernel_ext_info_size=%zu.", ext_info.size()); | |||
| GE_CHK_STATUS_RET(ext_handle->UpdateExecuteMode(true), "UpdateExecuteMode failed."); | |||
| GELOGD("Update aicpu_task ext_info bit_map execute mode to 1."); | |||
| auto rt_ret = rtMalloc(&ext_info_addr_, ext_handle->GetExtInfoLen(), RT_MEMORY_HBM); | |||
| GE_IF_BOOL_EXEC(rt_ret != RT_ERROR_NONE, | |||
| GELOGE(RT_FAILED, "rtMalloc ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret);) | |||
| rt_ret = rtMemcpy(ext_info_addr_, ext_handle->GetExtInfoLen(), ext_handle->GetExtInfo(), | |||
| ext_handle->GetExtInfoLen(), RT_MEMCPY_HOST_TO_DEVICE); | |||
| GE_IF_BOOL_EXEC(rt_ret != RT_ERROR_NONE, | |||
| GELOGE(RT_FAILED, "rtMemcpy ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret);) | |||
| return SUCCESS; | |||
| } | |||
| Status KernelExTaskInfo::Init(const domi::TaskDef &task_def, DavinciModel *davinci_model) { | |||
| GELOGI("KernelExTaskInfo Init Start."); | |||
| GE_CHECK_NOTNULL(davinci_model); | |||
| @@ -63,16 +97,8 @@ Status KernelExTaskInfo::Init(const domi::TaskDef &task_def, DavinciModel *davin | |||
| } | |||
| const auto &ext_info = kernel_ex_def.kernel_ext_info(); | |||
| if (!ext_info.empty()) { | |||
| auto rt_ret = rtMalloc(&ext_info_addr_, ext_info.size(), RT_MEMORY_HBM); | |||
| GE_IF_BOOL_EXEC(rt_ret != RT_ERROR_NONE, | |||
| GELOGE(RT_FAILED, "rtMalloc ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret);) | |||
| rt_ret = rtMemcpy(ext_info_addr_, ext_info.size(), ext_info.c_str(), ext_info.size(), RT_MEMCPY_HOST_TO_DEVICE); | |||
| GE_IF_BOOL_EXEC(rt_ret != RT_ERROR_NONE, | |||
| GELOGE(RT_FAILED, "rtMemcpy ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret);) | |||
| } | |||
| GE_CHK_STATUS_RET(InitTaskExtInfo(ext_info, op_desc), | |||
| "Init aicpu tf_task ext info failed, ext_info size=%zu", ext_info.size()); | |||
| GELOGI("Node[%s] type[%s] kernel_ext_info size=%zu, ext_info_addr_=%p", op_desc->GetName().c_str(), | |||
| op_desc->GetType().c_str(), ext_info.size(), ext_info_addr_); | |||
| @@ -62,6 +62,7 @@ class KernelExTaskInfo : public TaskInfo { | |||
| void SetIoAddrs(const OpDescPtr &op_desc); | |||
| void InitDumpTask(void *addr, const OpDescPtr &op_desc); | |||
| Status InitTaskExtInfo(const std::string &ext_info, const OpDescPtr &op_desc); | |||
| uint32_t task_id_; | |||
| uint32_t stream_id_; | |||
| @@ -32,6 +32,8 @@ | |||
| #include "super_kernel/super_kernel.h" | |||
| #include "super_kernel/super_kernel_factory.h" | |||
| #include "cce/aicpu_engine_struct.h" | |||
| #include "hybrid/node_executor/aicpu/aicpu_ext_info.h" | |||
| #include "framework/common/debug/log.h" | |||
| namespace { | |||
| const uint8_t kL2LoadToDdr = 1; | |||
| @@ -964,39 +966,32 @@ Status KernelTaskInfo::InitAicpuTaskExtInfo(const std::string &ext_info) { | |||
| return SUCCESS; | |||
| } | |||
| std::unique_ptr<uint8_t[]> copy_ext_info; | |||
| copy_ext_info.reset(new(std::nothrow)uint8_t[ext_info.size()]); | |||
| GE_CHECK_NOTNULL(copy_ext_info); | |||
| auto sec_ret = memcpy_s(copy_ext_info.get(), ext_info.size(), ext_info.c_str(), ext_info.size()); | |||
| if (sec_ret != EOK) { | |||
| GELOGE(FAILED, "memcpy failed, ret: %d", sec_ret); | |||
| return FAILED; | |||
| } | |||
| auto ext_info_data = copy_ext_info.get(); | |||
| size_t offset = 0; | |||
| while (offset + sizeof(aicpu::FWKAdapter::ExtInfo) <= ext_info.size()) { | |||
| auto aicpu_ext_info = reinterpret_cast<aicpu::FWKAdapter::ExtInfo *>(ext_info_data + offset); | |||
| GELOGD("Ext infoType=%d, infoLen=%u.", aicpu_ext_info->infoType, aicpu_ext_info->infoLen); | |||
| if (aicpu_ext_info->infoType == aicpu::FWKAdapter::FWK_ADPT_EXT_SESSION_INFO) { | |||
| GE_CHK_BOOL_RET_STATUS(aicpu_ext_info->infoLen == sizeof(SessionInfo), PARAM_INVALID, | |||
| "Parse ext session info failed as infoLen must be %zu but %u.", | |||
| sizeof(SessionInfo), aicpu_ext_info->infoLen); | |||
| SessionInfo *session_info = reinterpret_cast<SessionInfo *>(aicpu_ext_info->infoMsg); | |||
| session_info->sessionId = davinci_model_->GetSessionId(); | |||
| session_info->sessFlag = true; | |||
| GELOGD("Update aicpu_task ext_info session_info session_id is %lu", session_info->sessionId); | |||
| } | |||
| offset += sizeof(aicpu::FWKAdapter::ExtInfo); | |||
| offset += aicpu_ext_info->infoLen; | |||
| } | |||
| auto rt_ret = rtMalloc(&aicpu_ext_info_addr_, ext_info.size(), RT_MEMORY_HBM); | |||
| int32_t unknown_shape_type_val = 0; | |||
| (void) AttrUtils::GetInt(op_desc_, ::ge::ATTR_NAME_UNKNOWN_SHAPE_TYPE, unknown_shape_type_val); | |||
| UnknowShapeOpType unknown_type = static_cast<UnknowShapeOpType>(unknown_shape_type_val); | |||
| uint32_t num_inputs = op_desc_->GetInputsSize(); | |||
| uint32_t num_outputs = op_desc_->GetOutputsSize(); | |||
| std::unique_ptr<ge::hybrid::AicpuExtInfoHandler> ext_handle( | |||
| new(std::nothrow) ::ge::hybrid::AicpuExtInfoHandler(op_desc_->GetName(), | |||
| num_inputs, | |||
| num_outputs, | |||
| unknown_type)); | |||
| GE_CHK_BOOL_RET_STATUS(ext_handle != nullptr, FAILED, "Malloc aicpu_ext_handle mem failed!"); | |||
| GE_CHK_STATUS_RET(ext_handle->Parse(ext_info), | |||
| "Parse kernel ext info failed, kernel_ext_info_size=%zu.", ext_info.size()); | |||
| GE_CHK_STATUS_RET(ext_handle->UpdateSessionInfoSessionId(davinci_model_->GetSessionId()), | |||
| "Update session info session id failed."); | |||
| GELOGD("Update aicpu_task ext_info session_info session_id is %lu", davinci_model_->GetSessionId()); | |||
| GE_CHK_STATUS_RET(ext_handle->UpdateExecuteMode(true), "UpdateExecuteMode failed."); | |||
| GELOGD("Update aicpu_task ext_info bit_map execute mode to 1."); | |||
| auto rt_ret = rtMalloc(&aicpu_ext_info_addr_, ext_handle->GetExtInfoLen(), RT_MEMORY_HBM); | |||
| if (rt_ret != RT_ERROR_NONE) { | |||
| GELOGE(RT_FAILED, "rtMalloc ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret); | |||
| } | |||
| rt_ret = rtMemcpy(aicpu_ext_info_addr_, ext_info.size(), ext_info_data, ext_info.size(), RT_MEMCPY_HOST_TO_DEVICE); | |||
| rt_ret = rtMemcpy(aicpu_ext_info_addr_, ext_handle->GetExtInfoLen(), ext_handle->GetExtInfo(), | |||
| ext_handle->GetExtInfoLen(), RT_MEMCPY_HOST_TO_DEVICE); | |||
| if (rt_ret != RT_ERROR_NONE) { | |||
| GELOGE(RT_FAILED, "rtMemcpy ext_info error: 0x%X, size=%zu", rt_ret, ext_info.size()); | |||
| return RT_ERROR_TO_GE_STATUS(rt_ret); | |||
| @@ -92,7 +92,7 @@ | |||
| #include "graph/passes/unused_args_clean_pass.h" | |||
| #include "graph/passes/global_step_insert_pass.h" | |||
| #include "graph/passes/memcpy_addr_async_pass.h" | |||
| #include "graph/passes/hccl_memcpy_pass.h" | |||
| #include "graph/passes/hccl_continuous_memcpy_pass.h" | |||
| #include "graph/build/label_allocator.h" | |||
| #include "graph/utils/tensor_adapter.h" | |||
| #include "inc/pass_manager.h" | |||
| @@ -2151,8 +2151,6 @@ Status GraphManager::OptimizeStage1(ge::ComputeGraphPtr &compute_graph) { | |||
| new (std::nothrow) TransOpWithoutReshapeFusionPass)) | |||
| GE_CHK_STATUS_RET(after_merge_passes.AddPass("OptimizeStage1_1::TransOpBreadthFusionPass", | |||
| new (std::nothrow) TransOpBreadthFusionPass)) | |||
| GE_CHK_STATUS_RET( | |||
| after_merge_passes.AddPass("OptimizeStage1_1::HcclMemcpyPass", new (std::nothrow) HcclMemcpyPass)); | |||
| GE_TIMESTAMP_START(after_merge_passes); | |||
| auto ret = after_merge_passes.Run(compute_graph); | |||
| @@ -2268,6 +2266,9 @@ Status GraphManager::OptimizeStage2(ge::ComputeGraphPtr &compute_graph) { | |||
| GE_CHK_STATUS_RET(after_merge_passes.AddPass("OptimizeStage2::AfterMergePasses::LinkGenMaskNodesPass", | |||
| new (std::nothrow) | |||
| LinkGenMaskNodesPass(options_.stream_max_parallel_num))); | |||
| GE_CHK_STATUS_RET( | |||
| after_merge_passes.AddPass("OptimizeStage2::HcclContinuousMemcpyPass", | |||
| new (std::nothrow) HcclContinuousMemcpyPass)); | |||
| GE_TIMESTAMP_START(after_merge_passes); | |||
| auto ret = after_merge_passes.Run(compute_graph); | |||
| @@ -0,0 +1,411 @@ | |||
| /** | |||
| * Copyright 2020 Huawei Technologies Co., Ltd | |||
| * | |||
| * Licensed under the Apache License, Version 2.0 (the "License"); | |||
| * you may not use this file except in compliance with the License. | |||
| * You may obtain a copy of the License at | |||
| * | |||
| * http://www.apache.org/licenses/LICENSE-2.0 | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, software | |||
| * distributed under the License is distributed on an "AS IS" BASIS, | |||
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| * See the License for the specific language governing permissions and | |||
| * limitations under the License. | |||
| */ | |||
| #include "graph/passes/hccl_continuous_memcpy_pass.h" | |||
| #include <string> | |||
| #include "common/debug/log.h" | |||
| #include "framework/common/debug/ge_log.h" | |||
| #include "common/ge_inner_error_codes.h" | |||
| #include "common/ge/ge_util.h" | |||
| #include "framework/common/types.h" | |||
| #include "graph/utils/graph_utils.h" | |||
| namespace { | |||
| const int kAnchorNum = 0; | |||
| const int32_t kAnchorAssignRefIndex = 0; | |||
| const int32_t kAnchorAssignValueIndex = 1; | |||
| } // namespace | |||
| namespace ge { | |||
| Status HcclContinuousMemcpyPass::Run(ge::ComputeGraphPtr graph) { | |||
| GE_CHECK_NOTNULL(graph); | |||
| for (const auto &node : graph->GetDirectNode()) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| if (op_desc == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "node has no op_desc, node_name : %s.", node->GetName().c_str()); | |||
| return INTERNAL_ERROR; | |||
| } | |||
| Status ret = ContinuousInputProcess(graph, node); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "failed ProcessBroadcastMemcpy, node_name:%s.", node->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| ret = P2pmemInputProcess(graph, node); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "failed P2pmemInputProcess, node_name:%s.", node->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| // If broadcast input size is bigger than 1, and input from variable, | |||
| // cause by broadcast input memory should be continuous, | |||
| // another featuremap mem will be allocated for broadcast input. | |||
| // In this condition, move data from variable mem to broadcast input featuremap mem will be executed each step. | |||
| // In order to avoid move action out of model, use memcpy node instead of move action code. | |||
| Status HcclContinuousMemcpyPass::ContinuousInputProcess(const ComputeGraphPtr &graph, const NodePtr node) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| bool is_input_continuous = false; | |||
| (void)ge::AttrUtils::GetBool(op_desc, ATTR_NAME_CONTINUOUS_INPUT, is_input_continuous); | |||
| if (is_input_continuous && op_desc->GetInputsSize() > 1) { | |||
| GELOGI("continuous input op is:%s.", op_desc->GetName().c_str()); | |||
| // if input size bigger than one, insert memcpy between var data for support continous mem alloc | |||
| for (auto &hccl_in_anchor : node->GetAllInDataAnchors()) { | |||
| if (hccl_in_anchor == nullptr) { | |||
| continue; | |||
| } | |||
| auto src_out_anchor = hccl_in_anchor->GetPeerOutAnchor(); | |||
| if (src_out_anchor == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "hcom op input has no peer anchor, node_name:%s", node->GetName().c_str()); | |||
| return INTERNAL_ERROR; | |||
| } | |||
| if (IsDataNode(src_out_anchor->GetOwnerNode()->GetType())) { | |||
| Status ret = ModifyEdgeConnection(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Failed to modify the connection."); | |||
| return ret; | |||
| } | |||
| } | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| // if input is var type, and node input need p2p mem, then memcpy should be insert between the two | |||
| Status HcclContinuousMemcpyPass::P2pmemInputProcess(const ComputeGraphPtr &graph, const NodePtr node) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| vector<int64_t> input_memory_types; | |||
| (void) ge::AttrUtils::GetListInt(op_desc, ATTR_NAME_INPUT_MEM_TYPE_LIST, input_memory_types); | |||
| if (input_memory_types.empty()) { | |||
| return SUCCESS; | |||
| } | |||
| for (uint32_t index = 0; index < input_memory_types.size() && index < op_desc->GetInputsSize(); index++) { | |||
| if (input_memory_types[index] != RT_MEMORY_P2P_DDR) { | |||
| continue; | |||
| } | |||
| GELOGD("p2p input op is:%s.", op_desc->GetName().c_str()); | |||
| auto hccl_in_anchor = node->GetInDataAnchor(index); | |||
| if (hccl_in_anchor == nullptr) { | |||
| continue; | |||
| } | |||
| auto src_out_anchor = hccl_in_anchor->GetPeerOutAnchor(); | |||
| if (src_out_anchor == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "hcom op input has no peer anchor, node_name:%s", node->GetName().c_str()); | |||
| return INTERNAL_ERROR; | |||
| } | |||
| if (IsDataNode(src_out_anchor->GetOwnerNode()->GetType())) { | |||
| Status ret = ModifyEdgeConnection(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Failed to modify the connection."); | |||
| return ret; | |||
| } | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| bool HcclContinuousMemcpyPass::IsDataNode(const std::string& node_type) { | |||
| return (node_type == CONSTANTOP) || (node_type == VARIABLE) || (node_type == DATA) || (node_type == CONSTANT); | |||
| } | |||
| /// | |||
| /// @brief Add Identity Node | |||
| /// @param [in] ge::ComputeGraphPtr graph | |||
| /// @param [in] ge::OutDataAnchorPtr in_node | |||
| /// @return ge::NodePtr | |||
| /// | |||
| NodePtr HcclContinuousMemcpyPass::CreateIdentityNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor) { | |||
| GE_CHECK_NOTNULL_EXEC(graph, return nullptr); | |||
| NodePtr pre_node = out_data_anchor->GetOwnerNode(); | |||
| OpDescPtr pre_op_desc = pre_node->GetOpDesc(); | |||
| if (pre_op_desc == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "OpDesc of pre node is invalid."); | |||
| return nullptr; | |||
| } | |||
| std::string node_name = pre_node->GetName() + "_" + IDENTITY; | |||
| node_name = CheckDuplicateName(node_name); | |||
| OpDescPtr op_desc = MakeShared<OpDesc>(node_name.c_str(), IDENTITY); | |||
| if (op_desc == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "Create Identity op: MakeShared op_desc fail."); | |||
| return nullptr; | |||
| } | |||
| GELOGI("Create Identity op:%s.", op_desc->GetName().c_str()); | |||
| graphStatus ret = op_desc->AddInputDesc("x", pre_op_desc->GetOutputDesc(out_data_anchor->GetIdx())); | |||
| if (ret != GRAPH_SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Create Identity op: add input desc fail."); | |||
| return nullptr; | |||
| } | |||
| ret = op_desc->AddOutputDesc("y", pre_op_desc->GetOutputDesc(out_data_anchor->GetIdx())); | |||
| if (ret != GRAPH_SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Create Identity op: add output desc fail."); | |||
| return nullptr; | |||
| } | |||
| // because history reason ,this pass can not do work after constant fold so mark it | |||
| (void)AttrUtils::SetBool(op_desc, ATTR_NO_NEED_CONSTANT_FOLDING, false); | |||
| NodePtr memcpy_node = graph->AddNode(op_desc); | |||
| if (memcpy_node == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "Insert Identity node fail."); | |||
| return nullptr; | |||
| } | |||
| return memcpy_node; | |||
| } | |||
| /// | |||
| /// @brief Check duplicate node_name | |||
| /// @param [in] std::string& node_name | |||
| /// @return std::string | |||
| /// | |||
| std::string HcclContinuousMemcpyPass::CheckDuplicateName(const std::string &node_name) { | |||
| std::string tmp_name = node_name; | |||
| auto iter = node_num_map_.find(tmp_name); | |||
| if (iter != node_num_map_.end()) { | |||
| tmp_name = tmp_name + "_" + std::to_string(iter->second); | |||
| (iter->second)++; | |||
| } else { | |||
| node_num_map_[tmp_name] = 1; | |||
| } | |||
| return tmp_name; | |||
| } | |||
| /// | |||
| /// @brief Modify edge connection | |||
| /// @param [in] ComputeGraphPtr graph | |||
| /// @param [in] OutDataAnchorPtr src_out_anchor | |||
| /// @param [in] InDataAnchorPtr hccl_in_anchor | |||
| /// @return status | |||
| /// | |||
| Status HcclContinuousMemcpyPass::ModifyEdgeConnection(const ComputeGraphPtr &graph, const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor) { | |||
| GE_CHECK_NOTNULL(src_out_anchor->GetOwnerNode()); | |||
| GE_CHECK_NOTNULL(hccl_in_anchor->GetOwnerNode()); | |||
| Status ret = InsertIdentityBeforeHccl(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "add identity failed, var_node:%s, hccl_node:%s.", | |||
| src_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| ret = InsertAssignAfterBroadcastIfNeed(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "add assign failed, var_node:%s, hccl_node:%s.", | |||
| src_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| /// | |||
| /// @brief Insert Identity node Between Hccl node and variable | |||
| /// @param [in] ComputeGraphPtr graph | |||
| /// @param [in] OutDataAnchorPtr src_out_anchor | |||
| /// @param [in] InDataAnchorPtr hccl_in_anchor | |||
| /// @return status | |||
| /// | |||
| Status HcclContinuousMemcpyPass::InsertIdentityBeforeHccl(const ComputeGraphPtr &graph, const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor) { | |||
| GELOGI("Between op %s and op %s need insert memcpy async op.", src_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| NodePtr memcpy_node = CreateIdentityNode(graph, src_out_anchor); | |||
| GE_CHECK_NOTNULL(memcpy_node); | |||
| Status ret1 = src_out_anchor->Unlink(hccl_in_anchor); | |||
| if (ret1 != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s Unlink anchor %s fail.", src_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| auto out_data_anchor_0 = memcpy_node->GetOutDataAnchor(kAnchorNum); | |||
| GE_CHECK_NOTNULL(out_data_anchor_0); | |||
| ret1 = out_data_anchor_0->LinkTo(hccl_in_anchor); | |||
| if (ret1 != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link anchor %s fail.", memcpy_node->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| Status ret = src_out_anchor->LinkTo(memcpy_node->GetInDataAnchor(kAnchorNum)); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link anchor %s fail.", src_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| memcpy_node->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| /// | |||
| /// @brief Insert assign node after broadcast node and variable to refresh variable data | |||
| /// @param [in] ComputeGraphPtr graph | |||
| /// @param [in] OutDataAnchorPtr var_out_anchor | |||
| /// @param [in] InDataAnchorPtr hccl_in_anchor | |||
| /// @return status | |||
| /// | |||
| Status HcclContinuousMemcpyPass::InsertAssignAfterBroadcastIfNeed(const ComputeGraphPtr &graph, | |||
| const OutDataAnchorPtr &var_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor) { | |||
| if (hccl_in_anchor->GetOwnerNode()->GetType() != HCOMBROADCAST) { | |||
| GELOGD("%s not broadcast, no need to insert assign node", hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return SUCCESS; | |||
| } | |||
| if (var_out_anchor->GetOwnerNode()->GetType() != VARIABLE) { | |||
| GELOGD("%s not variable, no need to insert assign node", var_out_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return SUCCESS; | |||
| } | |||
| GELOGI("after op %s and op %s need insert assign op.", var_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| hccl_in_anchor->GetOwnerNode()->GetName().c_str()); | |||
| for (auto peer_in_anchor : var_out_anchor->GetPeerInDataAnchors()) { | |||
| if (peer_in_anchor->GetOwnerNode()->GetType() == ASSIGN) { | |||
| GELOGD("variable %s out assign node is exist.", var_out_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return SUCCESS; | |||
| } | |||
| } | |||
| NodePtr assign_node = CreateAssignNode(graph, var_out_anchor); | |||
| GE_CHECK_NOTNULL(assign_node); | |||
| OutDataAnchorPtr hccl_out_anchor = hccl_in_anchor->GetOwnerNode()->GetOutDataAnchor(hccl_in_anchor->GetIdx()); | |||
| GE_CHECK_NOTNULL(hccl_out_anchor); | |||
| Status ret = hccl_out_anchor->LinkTo(assign_node->GetInDataAnchor(kAnchorAssignValueIndex)); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link anchor %s fail.", hccl_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| assign_node->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| ret = var_out_anchor->LinkTo(assign_node->GetInDataAnchor(kAnchorAssignRefIndex)); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link anchor %s fail.", var_out_anchor->GetOwnerNode()->GetName().c_str(), | |||
| assign_node->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| // add control edge between assign node and node after broadcast node | |||
| OutControlAnchorPtr assign_out_control_anchor = assign_node->GetOutControlAnchor(); | |||
| GE_CHECK_NOTNULL(assign_out_control_anchor); | |||
| for (auto in_data_anchor : hccl_out_anchor->GetPeerInDataAnchors()) { | |||
| if (in_data_anchor->GetOwnerNode()->GetName() == assign_node->GetName()) { | |||
| continue; | |||
| } | |||
| ret = assign_out_control_anchor->LinkTo(in_data_anchor->GetOwnerNode()->GetInControlAnchor()); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link control anchor %s fail.", | |||
| assign_out_control_anchor->GetOwnerNode()->GetName().c_str(), | |||
| in_data_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| } | |||
| for (auto in_control_anchor : hccl_out_anchor->GetOwnerNode()->GetOutControlAnchor()->GetPeerInControlAnchors()) { | |||
| if (in_control_anchor->GetOwnerNode()->GetName() == assign_node->GetName()) { | |||
| continue; | |||
| } | |||
| ret = assign_out_control_anchor->LinkTo(in_control_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "The op %s link control anchor %s fail.", | |||
| assign_out_control_anchor->GetOwnerNode()->GetName().c_str(), | |||
| in_control_anchor->GetOwnerNode()->GetName().c_str()); | |||
| return FAILED; | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| /// | |||
| /// @brief create assign Node, add to graph | |||
| /// @param [in] ge::ComputeGraphPtr graph | |||
| /// @param [in] ge::OutDataAnchorPtr variable node out anchor | |||
| /// @return ge::NodePtr | |||
| /// | |||
| NodePtr HcclContinuousMemcpyPass::CreateAssignNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor) { | |||
| GE_CHECK_NOTNULL_EXEC(graph , return nullptr); | |||
| NodePtr pre_node = out_data_anchor->GetOwnerNode(); | |||
| OpDescPtr pre_op_desc = pre_node->GetOpDesc(); | |||
| if (pre_op_desc == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "OpDesc of pre node is invalid."); | |||
| return nullptr; | |||
| } | |||
| std::string node_name = pre_node->GetName() + "_" + ASSIGN; | |||
| node_name = CheckDuplicateName(node_name); | |||
| OpDescPtr op_desc = MakeShared<OpDesc>(node_name.c_str(), ASSIGN); | |||
| if (op_desc == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "Create Assign op: MakeShared op_desc fail."); | |||
| return nullptr; | |||
| } | |||
| GELOGI("Create Assign op:%s.", op_desc->GetName().c_str()); | |||
| graphStatus ret = op_desc->AddInputDesc("ref", pre_op_desc->GetOutputDesc(out_data_anchor->GetIdx())); | |||
| if (ret != GRAPH_SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Create Assign op: add ref input desc fail."); | |||
| return nullptr; | |||
| } | |||
| ret = op_desc->AddInputDesc("value", pre_op_desc->GetOutputDesc(out_data_anchor->GetIdx())); | |||
| if (ret != GRAPH_SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Create Assign op: add value input desc fail."); | |||
| return nullptr; | |||
| } | |||
| ret = op_desc->AddOutputDesc("ref", pre_op_desc->GetOutputDesc(out_data_anchor->GetIdx())); | |||
| if (ret != GRAPH_SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Create Assign op: add output desc fail."); | |||
| return nullptr; | |||
| } | |||
| NodePtr assign_node = graph->AddNode(op_desc); | |||
| if (assign_node == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "Insert Identity node fail."); | |||
| return nullptr; | |||
| } | |||
| return assign_node; | |||
| } | |||
| /// | |||
| /// @brief Clear Status, used for subgraph pass | |||
| /// @return SUCCESS | |||
| /// | |||
| Status HcclContinuousMemcpyPass::ClearStatus() { | |||
| node_num_map_.clear(); | |||
| return SUCCESS; | |||
| } | |||
| } // namespace ge | |||
| @@ -0,0 +1,59 @@ | |||
| /** | |||
| * Copyright 2020 Huawei Technologies Co., Ltd | |||
| * | |||
| * Licensed under the Apache License, Version 2.0 (the "License"); | |||
| * you may not use this file except in compliance with the License. | |||
| * You may obtain a copy of the License at | |||
| * | |||
| * http://www.apache.org/licenses/LICENSE-2.0 | |||
| * | |||
| * Unless required by applicable law or agreed to in writing, software | |||
| * distributed under the License is distributed on an "AS IS" BASIS, | |||
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||
| * See the License for the specific language governing permissions and | |||
| * limitations under the License. | |||
| */ | |||
| #ifndef GE_GRAPH_PASSES_HCCL_CONTINUOUS_MEMCPY_PASS_H_ | |||
| #define GE_GRAPH_PASSES_HCCL_CONTINUOUS_MEMCPY_PASS_H_ | |||
| #include <string> | |||
| #include <unordered_map> | |||
| #include "graph/graph.h" | |||
| #include "inc/graph_pass.h" | |||
| namespace ge { | |||
| class HcclContinuousMemcpyPass : public GraphPass { | |||
| public: | |||
| Status Run(ge::ComputeGraphPtr graph); | |||
| Status ClearStatus() override; | |||
| private: | |||
| NodePtr CreateIdentityNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor); | |||
| NodePtr CreateAssignNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor); | |||
| std::string CheckDuplicateName(const std::string &node_name); | |||
| Status ModifyEdgeConnection(const ComputeGraphPtr &graph, const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor); | |||
| Status InsertIdentityBeforeHccl(const ComputeGraphPtr &graph, const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor); | |||
| Status InsertAssignAfterBroadcastIfNeed(const ComputeGraphPtr &graph, | |||
| const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor); | |||
| Status ContinuousInputProcess(const ComputeGraphPtr &graph, const NodePtr node); | |||
| Status P2pmemInputProcess(const ComputeGraphPtr &graph, const NodePtr node); | |||
| bool IsDataNode(const std::string& node_type); | |||
| std::unordered_map<std::string, uint32_t> node_num_map_; | |||
| }; | |||
| } // namespace ge | |||
| #endif // GE_GRAPH_PASSES_HCCL_MEMCPY_PASS_H_ | |||
| @@ -34,7 +34,7 @@ const char *const kInputMutable = "_input_mutable"; | |||
| } // namespace | |||
| namespace ge { | |||
| Status HcclMemcpyPass::Run(ge::ComputeGraphPtr graph) { | |||
| GE_IF_BOOL_EXEC(graph == nullptr, GELOGE(PARAM_INVALID, "param [graph] must not be null."); return PARAM_INVALID); | |||
| GE_CHECK_NOTNULL(graph); | |||
| for (const auto &node : graph->GetDirectNode()) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| if (op_desc == nullptr) { | |||
| @@ -42,24 +42,11 @@ Status HcclMemcpyPass::Run(ge::ComputeGraphPtr graph) { | |||
| return INTERNAL_ERROR; | |||
| } | |||
| Status ret = ContinuousInputProcess(graph, node); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "failed ProcessBroadcastMemcpy, node_name:%s.", node->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| ret = MutableInputProcess(graph, node); | |||
| Status ret = MutableInputProcess(graph, node); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "failed MutableInputProcess, node_name:%s.", node->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| ret = P2pmemInputProcess(graph, node); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "failed P2pmemInputProcess, node_name:%s.", node->GetName().c_str()); | |||
| return ret; | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| @@ -114,80 +101,6 @@ Status HcclMemcpyPass::MutableInputProcess(const ComputeGraphPtr &graph, const N | |||
| return SUCCESS; | |||
| } | |||
| // If broadcast input size is bigger than 1, and input from variable, | |||
| // cause by broadcast input memory should be continuous, | |||
| // another featuremap mem will be allocated for broadcast input. | |||
| // In this condition, move data from variable mem to broadcast input featuremap mem will be executed each step. | |||
| // In order to avoid move action out of model, use memcpy node instead of move action code. | |||
| Status HcclMemcpyPass::ContinuousInputProcess(const ComputeGraphPtr &graph, const NodePtr node) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| bool is_input_continuous = false; | |||
| (void)ge::AttrUtils::GetBool(op_desc, ATTR_NAME_CONTINUOUS_INPUT, is_input_continuous); | |||
| if (is_input_continuous && op_desc->GetInputsSize() > 1) { | |||
| GELOGI("continuous input op is:%s.", op_desc->GetName().c_str()); | |||
| // if input size bigger than one, insert memcpy between var data for support continous mem alloc | |||
| for (auto &hccl_in_anchor : node->GetAllInDataAnchors()) { | |||
| if (hccl_in_anchor == nullptr) { | |||
| continue; | |||
| } | |||
| auto src_out_anchor = hccl_in_anchor->GetPeerOutAnchor(); | |||
| if (src_out_anchor == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "hcom op input has no peer anchor, node_name:%s", node->GetName().c_str()); | |||
| return INTERNAL_ERROR; | |||
| } | |||
| if (IsDataNode(src_out_anchor->GetOwnerNode()->GetType())) { | |||
| Status ret = ModifyEdgeConnection(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Failed to modify the connection."); | |||
| return ret; | |||
| } | |||
| } | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| // if input is var type, and node input need p2p mem, then memcpy should be insert between the two | |||
| Status HcclMemcpyPass::P2pmemInputProcess(const ComputeGraphPtr &graph, const NodePtr node) { | |||
| auto op_desc = node->GetOpDesc(); | |||
| vector<int64_t> input_memory_types; | |||
| (void) ge::AttrUtils::GetListInt(op_desc, ATTR_NAME_INPUT_MEM_TYPE_LIST, input_memory_types); | |||
| if (input_memory_types.empty()) { | |||
| return SUCCESS; | |||
| } | |||
| for (uint32_t index = 0; index < input_memory_types.size() && index < op_desc->GetInputsSize(); index++) { | |||
| if (input_memory_types[index] != RT_MEMORY_P2P_DDR) { | |||
| continue; | |||
| } | |||
| GELOGD("p2p input op is:%s.", op_desc->GetName().c_str()); | |||
| auto hccl_in_anchor = node->GetInDataAnchor(index); | |||
| if (hccl_in_anchor == nullptr) { | |||
| continue; | |||
| } | |||
| auto src_out_anchor = hccl_in_anchor->GetPeerOutAnchor(); | |||
| if (src_out_anchor == nullptr) { | |||
| GELOGE(INTERNAL_ERROR, "hcom op input has no peer anchor, node_name:%s", node->GetName().c_str()); | |||
| return INTERNAL_ERROR; | |||
| } | |||
| if (IsDataNode(src_out_anchor->GetOwnerNode()->GetType())) { | |||
| Status ret = ModifyEdgeConnection(graph, src_out_anchor, hccl_in_anchor); | |||
| if (ret != SUCCESS) { | |||
| GELOGE(INTERNAL_ERROR, "Failed to modify the connection."); | |||
| return ret; | |||
| } | |||
| } | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| bool HcclMemcpyPass::IsDataNode(const std::string& node_type) { | |||
| return (node_type == CONSTANTOP) || (node_type == VARIABLE) || (node_type == DATA) || (node_type == CONSTANT); | |||
| } | |||
| @@ -199,7 +112,7 @@ bool HcclMemcpyPass::IsDataNode(const std::string& node_type) { | |||
| /// @return ge::NodePtr | |||
| /// | |||
| NodePtr HcclMemcpyPass::CreateIdentityNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor) { | |||
| GE_IF_BOOL_EXEC(graph == nullptr, return nullptr); | |||
| GE_CHECK_NOTNULL_EXEC(graph, return nullptr); | |||
| NodePtr pre_node = out_data_anchor->GetOwnerNode(); | |||
| OpDescPtr pre_op_desc = pre_node->GetOpDesc(); | |||
| if (pre_op_desc == nullptr) { | |||
| @@ -413,7 +326,7 @@ Status HcclMemcpyPass::InsertAssignAfterBroadcastIfNeed(const ComputeGraphPtr &g | |||
| /// @return ge::NodePtr | |||
| /// | |||
| NodePtr HcclMemcpyPass::CreateAssignNode(const ComputeGraphPtr &graph, const OutDataAnchorPtr &out_data_anchor) { | |||
| GE_IF_BOOL_EXEC(graph == nullptr, return nullptr); | |||
| GE_CHECK_NOTNULL_EXEC(graph, return nullptr); | |||
| NodePtr pre_node = out_data_anchor->GetOwnerNode(); | |||
| OpDescPtr pre_op_desc = pre_node->GetOpDesc(); | |||
| if (pre_op_desc == nullptr) { | |||
| @@ -46,12 +46,8 @@ class HcclMemcpyPass : public GraphPass { | |||
| const OutDataAnchorPtr &src_out_anchor, | |||
| const InDataAnchorPtr &hccl_in_anchor); | |||
| Status ContinuousInputProcess(const ComputeGraphPtr &graph, const NodePtr node); | |||
| Status MutableInputProcess(const ComputeGraphPtr &graph, const NodePtr node); | |||
| Status P2pmemInputProcess(const ComputeGraphPtr &graph, const NodePtr node); | |||
| bool IsDataNode(const std::string& node_type); | |||
| std::unordered_map<std::string, uint32_t> node_num_map_; | |||
| @@ -48,6 +48,7 @@ | |||
| #include "graph/passes/enter_pass.h" | |||
| #include "graph/passes/for_pass.h" | |||
| #include "graph/passes/guarantee_const_pass.h" | |||
| #include "graph/passes/hccl_memcpy_pass.h" | |||
| #include "graph/passes/hccl_group_pass.h" | |||
| #include "graph/passes/identity_pass.h" | |||
| #include "graph/passes/infershape_pass.h" | |||
| @@ -1894,6 +1895,8 @@ Status GraphPrepare::PrepareOptimize() { | |||
| PassManager graph_pass; | |||
| try { | |||
| (void)graph_pass.AddPass("PrepareOptimize::PrunePass", new PrunePass); | |||
| // can't move to optimize1/2 directly, may cause more identity insert, cause CI fail | |||
| (void)graph_pass.AddPass("PrepareOptimize::HcclMemcpyPass", new HcclMemcpyPass); | |||
| } catch (std::bad_alloc &e) { | |||
| GELOGE(INTERNAL_ERROR, "Add pass failed, bad memory allocation occurs."); | |||
| return INTERNAL_ERROR; | |||
| @@ -64,6 +64,9 @@ Status AicpuExtInfoHandler::Parse(const std::string &ext_info) { | |||
| case aicpu::FWKAdapter::FWK_ADPT_EXT_SESSION_INFO: | |||
| GE_CHK_STATUS_RET(ParseExtSessionInfo(aicpu_ext_info), "Parse ext session info failed."); | |||
| break; | |||
| case aicpu::FWKAdapter::FWK_ADPT_EXT_BITMAP: | |||
| GE_CHK_STATUS_RET(ParseExtBitMap(aicpu_ext_info), "Parse ext bit map failed."); | |||
| break; | |||
| default: | |||
| GELOGD("Node[%s] ignore infoType=%d, infoLen=%u.", | |||
| node_name_.c_str(), aicpu_ext_info->infoType, aicpu_ext_info->infoLen); | |||
| @@ -140,6 +143,29 @@ Status AicpuExtInfoHandler::ParseExtSessionInfo(AicpuExtInfo *aicpu_ext_info) { | |||
| return SUCCESS; | |||
| } | |||
| Status AicpuExtInfoHandler::ParseExtBitMap(AicpuExtInfo *aicpu_ext_info) { | |||
| GE_CHK_BOOL_RET_STATUS(aicpu_ext_info->infoLen == sizeof(uint64_t), PARAM_INVALID, | |||
| "Node[%s] parse bit_map info failed as infoLen must be %zu but %u.", | |||
| node_name_.c_str(), sizeof(uint64_t), aicpu_ext_info->infoLen); | |||
| bit_map_ = reinterpret_cast<uint64_t *>(aicpu_ext_info->infoMsg); | |||
| GELOGI("Node[%s] bit_map info success infoLen=%u.", node_name_.c_str(), aicpu_ext_info->infoLen); | |||
| return SUCCESS; | |||
| } | |||
| Status AicpuExtInfoHandler::UpdateExecuteMode(bool flag) { | |||
| if (bit_map_ == nullptr) { | |||
| GELOGD("There is no bit_map in ext_info, no need update."); | |||
| return SUCCESS; | |||
| } | |||
| if (flag) { | |||
| *(bit_map_) |= 1; | |||
| } else { | |||
| *(bit_map_) &= ~1; | |||
| } | |||
| return SUCCESS; | |||
| } | |||
| Status AicpuExtInfoHandler::UpdateSessionInfo(uint64_t session_id, uint64_t kernel_id, bool sess_flag) { | |||
| if (session_info_ == nullptr) { | |||
| GELOGD("There is no session info in ext_info, no need update."); | |||
| @@ -57,6 +57,8 @@ class AicpuExtInfoHandler { | |||
| Status UpdateSessionInfoSessionId(uint64_t session_id); | |||
| Status UpdateExecuteMode(bool flag); | |||
| Status GetOutputShapeAndType(uint32_t output_index, GeShape &shape, DataType &data_type); | |||
| private: | |||
| @@ -65,6 +67,7 @@ class AicpuExtInfoHandler { | |||
| Status ParseExtInputShape(AicpuExtInfo *aicpu_ext_info); | |||
| Status ParseExtOutputShape(AicpuExtInfo *aicpu_ext_info); | |||
| Status ParseExtSessionInfo(AicpuExtInfo *aicpu_ext_info); | |||
| Status ParseExtBitMap(AicpuExtInfo *aicpu_ext_info); | |||
| static Status UpdateShapeAndType(const GeShape &shape, | |||
| DataType data_type, | |||
| @@ -80,6 +83,7 @@ class AicpuExtInfoHandler { | |||
| const uint32_t output_num_; | |||
| UnknowShapeOpType unknown_type_; | |||
| AicpuSessionInfo *session_info_ = nullptr; | |||
| uint64_t *bit_map_ = nullptr; | |||
| std::unique_ptr<uint8_t[]> ext_info_; | |||
| size_t ext_info_len_ = 0; | |||
| @@ -136,6 +136,7 @@ Status AicpuNodeTaskBase::UpdateExtInfo() { | |||
| return SUCCESS; | |||
| } | |||
| GE_CHK_STATUS_RET(aicpu_ext_handle_.UpdateExecuteMode(false), "UpdateExecuteMode failed."); | |||
| for (auto i = 0; i < node_item_->num_inputs; ++i) { | |||
| auto input_desc = node_item_->MutableInputDesc(i); | |||
| GE_CHECK_NOTNULL(input_desc); | |||
| @@ -373,6 +373,7 @@ Status AiCpuBaseTask::SetExtInfoAndType(const std::string &kernel_ext_info, uint | |||
| GE_CHK_STATUS_RET(aicpu_ext_handle_->UpdateSessionInfo(ULLONG_MAX, kernel_id, false), | |||
| "UpdateSessionInfo failed."); | |||
| GE_CHK_STATUS_RET(aicpu_ext_handle_->UpdateExecuteMode(true), "UpdateExecuteMode failed."); | |||
| GE_CHK_RT_RET(rtMalloc(&ext_info_addr_dev_, aicpu_ext_handle_->GetExtInfoLen(), RT_MEMORY_HBM)); | |||
| GE_CHK_RT_RET(rtMemcpy(ext_info_addr_dev_, aicpu_ext_handle_->GetExtInfoLen(), | |||
| @@ -404,13 +405,14 @@ Status AiCpuBaseTask::UpdateExtInfo(const std::vector<GeTensorDesc> &input_desc, | |||
| std::vector<GeTensorDesc> &output_desc, | |||
| rtStream_t stream) { | |||
| GELOGI("Update ext info begin, unknown_type=%d.", unknown_type_); | |||
| GE_CHECK_NOTNULL(aicpu_ext_handle_); | |||
| GE_CHK_STATUS_RET(aicpu_ext_handle_->UpdateExecuteMode(false), "UpdateExecuteMode failed."); | |||
| if (num_inputs_ == 0 && num_outputs_ == 0) { | |||
| GELOGI("No input and output, no need update ext info."); | |||
| return SUCCESS; | |||
| } | |||
| GE_CHECK_NOTNULL(aicpu_ext_handle_); | |||
| size_t non_const_index = 0; | |||
| for (size_t input_index = 0; input_index < num_inputs_; input_index++) { | |||
| if (input_index < input_is_const_.size() && input_is_const_[input_index]) { | |||
| @@ -243,6 +243,7 @@ set(COMMON_SRC_FILES | |||
| "${GE_CODE_DIR}/ge/graph/passes/cast_remove_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/transpose_transdata_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/hccl_memcpy_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/hccl_continuous_memcpy_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/flow_ctrl_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/global_step_insert_pass.cc" | |||
| "${GE_CODE_DIR}/ge/graph/passes/link_gen_mask_nodes_pass.cc" | |||
| @@ -137,4 +137,12 @@ TEST_F(UtestKernelExTaskInfo, kernel_ex_task_info_calculate_args) { | |||
| EXPECT_EQ(kernel_ex_task_info.CalculateArgs(task_def, &model), FAILED); | |||
| } | |||
| TEST_F(UtestKernelExTaskInfo, kernel_ex_task_ext_info) { | |||
| const string ext_info = {1, 1, 1, 1, 0, 0, 0, 0}; | |||
| const OpDescPtr op_desc = CreateOpDesc("FrameworkOp", "FrameworkOp"); | |||
| KernelExTaskInfo kernel_ex_task_info; | |||
| EXPECT_EQ(kernel_ex_task_info.InitTaskExtInfo(ext_info, op_desc), SUCCESS); | |||
| } | |||
| } // namespace ge | |||
| @@ -413,6 +413,9 @@ TEST_F(UtestKernelTaskInfo, init_kernel_taskInfo_with_aicpu_kernel_type_fail) { | |||
| // rtMemcpy -> RT_ERROR_INVALID_VALUE | |||
| EXPECT_EQ(kernel_task_info.Init(task_def, &model), SUCCESS); | |||
| const string ext_info = {1, 1, 1, 1, 0, 0, 0, 0}; | |||
| EXPECT_EQ(kernel_task_info.InitAicpuTaskExtInfo(ext_info), SUCCESS); | |||
| EXPECT_EQ(kernel_task_info.Distribute(), SUCCESS); | |||
| EXPECT_EQ(kernel_task_info.Release(), SUCCESS); | |||
| @@ -120,15 +120,15 @@ typedef struct tagKV { | |||
| } KeyValue; | |||
| typedef enum { | |||
| APPLICATION = 0, | |||
| SYSTEM | |||
| APPLICATION = 0, | |||
| SYSTEM | |||
| } ProcessType; | |||
| typedef struct { | |||
| ProcessType type; | |||
| unsigned int pid; | |||
| unsigned int deviceId; | |||
| char reserved[RESERVERD_LENGTH]; | |||
| ProcessType type; | |||
| unsigned int pid; | |||
| unsigned int deviceId; | |||
| char reserved[RESERVERD_LENGTH]; | |||
| } LogAttr; | |||
| /** | |||
| @@ -381,13 +381,13 @@ DLL_EXPORT void DlogFlush(void); | |||
| * @ingroup slog | |||
| * @brief Internal log interface, other modules are not allowed to call this interface | |||
| */ | |||
| void DlogErrorInner(int moduleId, const char *fmt, ...); | |||
| void DlogWarnInner(int moduleId, const char *fmt, ...); | |||
| void DlogInfoInner(int moduleId, const char *fmt, ...); | |||
| void DlogDebugInner(int moduleId, const char *fmt, ...); | |||
| void DlogEventInner(int moduleId, const char *fmt, ...); | |||
| void DlogInner(int moduleId, int level, const char *fmt, ...); | |||
| void DlogWithKVInner(int moduleId, int level, KeyValue *pstKVArray, int kvNum, const char *fmt, ...); | |||
| void DlogErrorInner(int moduleId, const char *fmt, ...) __attribute__((format(printf, 2, 3))); | |||
| void DlogWarnInner(int moduleId, const char *fmt, ...) __attribute__((format(printf, 2, 3))); | |||
| void DlogInfoInner(int moduleId, const char *fmt, ...) __attribute__((format(printf, 2, 3))); | |||
| void DlogDebugInner(int moduleId, const char *fmt, ...) __attribute__((format(printf, 2, 3))); | |||
| void DlogEventInner(int moduleId, const char *fmt, ...) __attribute__((format(printf, 2, 3))); | |||
| void DlogInner(int moduleId, int level, const char *fmt, ...) __attribute__((format(printf, 3, 4))); | |||
| void DlogWithKVInner(int moduleId, int level, KeyValue *pstKVArray, int kvNum, const char *fmt, ...) __attribute__((format(printf, 5, 6))); | |||
| #ifdef __cplusplus | |||
| #ifndef LOG_CPP | |||
| @@ -500,8 +500,8 @@ DLL_EXPORT void DlogFlushForC(void); | |||
| * @ingroup slog | |||
| * @brief Internal log interface, other modules are not allowed to call this interface | |||
| */ | |||
| void DlogInnerForC(int moduleId, int level, const char *fmt, ...); | |||
| void DlogWithKVInnerForC(int moduleId, int level, KeyValue *pstKVArray, int kvNum, const char *fmt, ...); | |||
| void DlogInnerForC(int moduleId, int level, const char *fmt, ...) __attribute__((format(printf, 3, 4))); | |||
| void DlogWithKVInnerForC(int moduleId, int level, KeyValue *pstKVArray, int kvNum, const char *fmt, ...) __attribute__((format(printf, 5, 6))); | |||
| #ifdef __cplusplus | |||
| } | |||