@@ -120,11 +120,11 @@ Status ModelHelper::SaveModelTbeKernel(std::shared_ptr<OmFileSaveHelper> &om_fil | |||||
TBEKernelStore tbe_kernel_store = ge_model->GetTBEKernelStore(); | TBEKernelStore tbe_kernel_store = ge_model->GetTBEKernelStore(); | ||||
GELOGD("TBE_KERNELS size is %zu", tbe_kernel_store.DataSize()); | GELOGD("TBE_KERNELS size is %zu", tbe_kernel_store.DataSize()); | ||||
if (tbe_kernel_store.DataSize() > 0) { | if (tbe_kernel_store.DataSize() > 0) { | ||||
GE_CHK_STATUS_RET(SaveModelPartition(om_file_save_helper, | |||||
ModelPartitionType::TBE_KERNELS, | |||||
ge_model->GetTBEKernelStore().Data(), | |||||
ge_model->GetTBEKernelStore().DataSize(), model_index), | |||||
"Add tbe kernel partition failed"); | |||||
GE_CHK_STATUS_RET( | |||||
SaveModelPartition(om_file_save_helper, ModelPartitionType::TBE_KERNELS, | |||||
ge_model->GetTBEKernelStore().Data(), | |||||
ge_model->GetTBEKernelStore().DataSize(), model_index), | |||||
"Add tbe kernel partition failed"); | |||||
} | } | ||||
// no need to check value, DATA->NetOutput | // no need to check value, DATA->NetOutput | ||||
(void)tbe_kernel_store.Load(tbe_kernel_store.Data(), tbe_kernel_store.DataSize()); | (void)tbe_kernel_store.Load(tbe_kernel_store.Data(), tbe_kernel_store.DataSize()); | ||||
@@ -2695,8 +2695,9 @@ Status DavinciModel::ReturnResult(uint32_t data_id, const bool rslt_flg, const b | |||||
is_getnext_sink_dynamic_ = true; | is_getnext_sink_dynamic_ = true; | ||||
cur_dynamic_dims_.clear(); | cur_dynamic_dims_.clear(); | ||||
cur_dynamic_dims_.resize(shape_of_cur_dynamic_dims_); | cur_dynamic_dims_.resize(shape_of_cur_dynamic_dims_); | ||||
GE_CHK_RT_RET(rtMemcpy(cur_dynamic_dims_.data(), shape_of_cur_dynamic_dims_ * sizeof(int64_t), | |||||
netoutput_last_input_addr_, netoutput_last_input_size_, RT_MEMCPY_DEVICE_TO_HOST)); | |||||
auto ret = rtMemcpy(cur_dynamic_dims_.data(), shape_of_cur_dynamic_dims_ * sizeof(int64_t), | |||||
netoutput_last_input_addr_, netoutput_last_input_size_, RT_MEMCPY_DEVICE_TO_HOST); | |||||
GE_CHK_RT_RET(ret); | |||||
} | } | ||||
GELOGD("Cur dynamic dims is %s.", formats::JoinToString(cur_dynamic_dims_).c_str()); | GELOGD("Cur dynamic dims is %s.", formats::JoinToString(cur_dynamic_dims_).c_str()); | ||||
if (GenOutputTensorInfo(op_desc, data_index, output_data, outputs) != SUCCESS) { | if (GenOutputTensorInfo(op_desc, data_index, output_data, outputs) != SUCCESS) { | ||||
@@ -1086,8 +1086,10 @@ Status ModelManager::GenSessionId(uint64_t &session_id) { | |||||
Status ModelManager::LoadModelOffline(uint32_t &model_id, const ModelData &model, shared_ptr<ModelListener> listener, | Status ModelManager::LoadModelOffline(uint32_t &model_id, const ModelData &model, shared_ptr<ModelListener> listener, | ||||
void *dev_ptr, size_t mem_size, void *weight_ptr, size_t weight_size) { | void *dev_ptr, size_t mem_size, void *weight_ptr, size_t weight_size) { | ||||
GE_CHK_BOOL_RET_STATUS(model.key.empty() || mmAccess2(model.key.c_str(), M_F_OK) == EN_OK, | GE_CHK_BOOL_RET_STATUS(model.key.empty() || mmAccess2(model.key.c_str(), M_F_OK) == EN_OK, | ||||
ACL_ERROR_GE_PARAM_INVALID, "input key file path %s is invalid, %s", | |||||
model.key.c_str(), strerror(errno)); | |||||
ACL_ERROR_GE_PARAM_INVALID, | |||||
"input key file path %s is invalid, %s", | |||||
model.key.c_str(), | |||||
strerror(errno)); | |||||
GenModelId(&model_id); | GenModelId(&model_id); | ||||
shared_ptr<DavinciModel> davinci_model = nullptr; | shared_ptr<DavinciModel> davinci_model = nullptr; | ||||
@@ -877,7 +877,7 @@ Status KernelTaskInfo::InitAicpuTask(uint32_t op_index, const domi::KernelDef &k | |||||
if (kernel_type_ == ccKernelType::CUST_AI_CPU) { | if (kernel_type_ == ccKernelType::CUST_AI_CPU) { | ||||
bool loaded = false; | bool loaded = false; | ||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc, so_name_, loaded), | GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc, so_name_, loaded), | ||||
"launch cust aicpu so failed"); | |||||
"launch cust aicpu so failed"); | |||||
} | } | ||||
// copy args to new host memory | // copy args to new host memory | ||||
@@ -646,7 +646,7 @@ Status AicpuNodeTask::Init(const HybridModel &model) { | |||||
if (kernel_type == ccKernelType::CUST_AI_CPU) { | if (kernel_type == ccKernelType::CUST_AI_CPU) { | ||||
bool loaded = false; | bool loaded = false; | ||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc, so_name, loaded), | GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc, so_name, loaded), | ||||
"load cust aicpu so failed."); | |||||
"load cust aicpu so failed."); | |||||
if (!loaded) { | if (!loaded) { | ||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LaunchCustAicpuSo(), "Launch cust aicpu so failed."); | GE_CHK_STATUS_RET(ModelManager::GetInstance()->LaunchCustAicpuSo(), "Launch cust aicpu so failed."); | ||||
} | } | ||||
@@ -64,7 +64,7 @@ Status AiCpuCCTaskBuilder::BuildTask(AiCpuCCTask &task, uint64_t kernel_id) { | |||||
task.dump_flag_ |= RT_KERNEL_CUSTOM_AICPU; | task.dump_flag_ |= RT_KERNEL_CUSTOM_AICPU; | ||||
bool loaded = false; | bool loaded = false; | ||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc_, so_name, loaded), | GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc_, so_name, loaded), | ||||
"launch cust aicpu so failed"); | |||||
"launch cust aicpu so failed"); | |||||
if (!loaded) { | if (!loaded) { | ||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LaunchCustAicpuSo(), "launch cust aicpu so failed."); | GE_CHK_STATUS_RET(ModelManager::GetInstance()->LaunchCustAicpuSo(), "launch cust aicpu so failed."); | ||||
} | } | ||||