Browse Source

Pre Merge pull request !1816 from guopeian/ge-taskInfo

pull/1816/MERGE
guopeian Gitee 4 years ago
parent
commit
00b3fa3a40
2 changed files with 130 additions and 57 deletions
  1. +103
    -32
      ge/hybrid/node_executor/aicpu/aicpu_node_executor.cc
  2. +27
    -25
      ge/hybrid/node_executor/aicpu/aicpu_node_executor.h

+ 103
- 32
ge/hybrid/node_executor/aicpu/aicpu_node_executor.cc View File

@@ -230,7 +230,7 @@ Status AicpuNodeTaskBase::ExecuteAsync(TaskContext &context, std::function<void(
return SUCCESS; return SUCCESS;
} }


Status AicpuTfNodeTask::InitForDependComputeTask() {
Status AicpuNodeTaskBase::InitForDependComputeTask(bool is_aicpu_kernel) {
if ((unknown_type_ != DEPEND_COMPUTE) || (node_item_->num_outputs == 0)) { if ((unknown_type_ != DEPEND_COMPUTE) || (node_item_->num_outputs == 0)) {
GELOGD("Node[%s] type[%s] unknown_type is %d, output num is %d.", GELOGD("Node[%s] type[%s] unknown_type is %d, output num is %d.",
node_name_.c_str(), node_item_->node_type.c_str(), unknown_type_, node_item_->num_outputs); node_name_.c_str(), node_item_->node_type.c_str(), unknown_type_, node_item_->num_outputs);
@@ -263,9 +263,15 @@ Status AicpuTfNodeTask::InitForDependComputeTask() {
node_name_.c_str(), copy_input_buf_len); node_name_.c_str(), copy_input_buf_len);


// copy task args buf // copy task args buf
GE_CHK_STATUS_RET(AllocTensorBuffer(sizeof(STR_FWK_OP_KERNEL), copy_task_args_buf_),
"[Alloc][TensorBuffer] failed for Node[%s] to copy task args, size=%zu",
node_name_.c_str(), sizeof(STR_FWK_OP_KERNEL));
if (is_aicpu_kernel) {
GE_CHK_STATUS_RET(AllocTensorBuffer(sizeof(aicpu::AicpuParamHead), copy_task_args_buf_),
"[Alloc][TensorBuffer] failed for Node[%s] to copy task args, size=%zu",
node_name_.c_str(), sizeof(aicpu::AicpuParamHead));
} else {
GE_CHK_STATUS_RET(AllocTensorBuffer(sizeof(STR_FWK_OP_KERNEL), copy_task_args_buf_),
"[Alloc][TensorBuffer] failed for Node[%s] to copy task args, size=%zu",
node_name_.c_str(), sizeof(STR_FWK_OP_KERNEL));
}


std::vector<uint64_t> copy_io_addr; std::vector<uint64_t> copy_io_addr;
copy_io_addr.emplace_back(reinterpret_cast<uintptr_t>(copy_input_release_flag_dev_->GetData())); copy_io_addr.emplace_back(reinterpret_cast<uintptr_t>(copy_input_release_flag_dev_->GetData()));
@@ -327,7 +333,7 @@ Status AicpuTfNodeTask::Init(const HybridModel &model) {
uint64_t ext_session_id = model.GetSessionId(); uint64_t ext_session_id = model.GetSessionId();
GE_CHK_STATUS_RET(InitExtInfo(kernel_ext_info, ext_session_id), "[Init][ExtInfo] failed for Node[%s].", GE_CHK_STATUS_RET(InitExtInfo(kernel_ext_info, ext_session_id), "[Init][ExtInfo] failed for Node[%s].",
node_name_.c_str()); node_name_.c_str());
GE_CHK_STATUS_RET(InitForDependComputeTask(), "[Init][DependComputeTask] failed for Node[%s].", node_name_.c_str());
GE_CHK_STATUS_RET(InitForDependComputeTask(false), "[Init][DependComputeTask] failed for Node[%s].", node_name_.c_str());


// build fwk_op_kernel. // build fwk_op_kernel.
GE_IF_BOOL_EXEC(sizeof(STR_FWK_OP_KERNEL) < kernel_ex_def.args_size(), GE_IF_BOOL_EXEC(sizeof(STR_FWK_OP_KERNEL) < kernel_ex_def.args_size(),
@@ -585,7 +591,8 @@ Status AicpuTfNodeTask::UpdateShapeAndDataByResultSummary(TaskContext &context)
return SUCCESS; return SUCCESS;
} }


Status AicpuTfNodeTask::UpdateIoAddr(TaskContext &context) {
Status AicpuTfNodeTask::
(TaskContext &context) {
vector<uint64_t> io_addrs; vector<uint64_t> io_addrs;
io_addrs.reserve(node_item_->num_inputs + node_item_->num_outputs); io_addrs.reserve(node_item_->num_inputs + node_item_->num_outputs);
for (auto i = 0; i < node_item_->num_inputs; ++i) { for (auto i = 0; i < node_item_->num_inputs; ++i) {
@@ -648,31 +655,10 @@ Status AicpuTfNodeTask::LaunchTask(TaskContext &context) {
return SUCCESS; return SUCCESS;
} }


Status AicpuTfNodeTask::TaskCallback(TaskContext &context) {
GELOGD("Node[%s] task callback start. is_dynamic=%s, unknown_type=%d.",
node_name_.c_str(), node_item_->is_dynamic ? "true" : "false", unknown_type_);
Status callback_ret = SUCCESS;
if (node_item_->is_dynamic) {
// check need update shape, call update shape.
if (unknown_type_ == DEPEND_SHAPE_RANGE) {
// check result
callback_ret = UpdateOutputShapeFromExtInfo(context);
} else if (unknown_type_ == DEPEND_COMPUTE) {
callback_ret = UpdateShapeAndDataByResultSummary(context);
}
}
GELOGD("Node[%s] task callback end.", node_name_.c_str());
return callback_ret;
}

Status AicpuNodeTask::Init(const HybridModel &model) { Status AicpuNodeTask::Init(const HybridModel &model) {
auto node_name = node_name_; auto node_name = node_name_;
GELOGD("Node[%s] init start.", node_name.c_str()); GELOGD("Node[%s] init start.", node_name.c_str());


GE_CHK_BOOL_RET_STATUS(unknown_type_ != DEPEND_COMPUTE, FAILED,
"[Check][Type]Node[%s] unknown type[%d] is depend compute, it's not supported now.",
node_name.c_str(), unknown_type_);

GE_CHK_BOOL_RET_STATUS(task_def_.has_kernel(), FAILED, GE_CHK_BOOL_RET_STATUS(task_def_.has_kernel(), FAILED,
"[Check][task_def_]Node[%s] task def does not has kernel.", node_name.c_str()); "[Check][task_def_]Node[%s] task def does not has kernel.", node_name.c_str());
auto &kernel_def = task_def_.kernel(); auto &kernel_def = task_def_.kernel();
@@ -761,6 +747,8 @@ Status AicpuNodeTask::Init(const HybridModel &model) {
uint64_t ext_session_id = model.GetSessionId(); uint64_t ext_session_id = model.GetSessionId();
GE_CHK_STATUS_RET(InitExtInfo(kernel_ext_info, ext_session_id), GE_CHK_STATUS_RET(InitExtInfo(kernel_ext_info, ext_session_id),
"[Init][ExtInfo] failed for Node[%s].", node_name.c_str()); "[Init][ExtInfo] failed for Node[%s].", node_name.c_str());
GE_CHK_STATUS_RET(InitForDependComputeTask(true),
"[Init][DependComputeTask] failed for Node[%s].", node_name_.c_str());


if (ext_info_addr_dev_ == nullptr) { if (ext_info_addr_dev_ == nullptr) {
aicpu_param_head->extInfoLength = 0; aicpu_param_head->extInfoLength = 0;
@@ -769,11 +757,65 @@ Status AicpuNodeTask::Init(const HybridModel &model) {
aicpu_param_head->extInfoLength = ext_info_addr_dev_->GetSize(); aicpu_param_head->extInfoLength = ext_info_addr_dev_->GetSize();
aicpu_param_head->extInfoAddr = reinterpret_cast<uintptr_t>(ext_info_addr_dev_->GetData()); aicpu_param_head->extInfoAddr = reinterpret_cast<uintptr_t>(ext_info_addr_dev_->GetData());
} }

auto task_defs = model.GetTaskDefs(node_item_->node);
GE_CHECK_NOTNULL(task_defs);
if (unknown_type_ == DEPEND_COMPUTE) {
GE_CHK_STATUS_RET_NOLOG(SetMemCopyTask((*task_defs)[1]));
}
GELOGD("Node[%s] init end.", node_name.c_str()); GELOGD("Node[%s] init end.", node_name.c_str());
return SUCCESS; return SUCCESS;
} }


Status AicpuNodeTask::SetMemCopyTask(const domi::TaskDef &task_def) {
if (node_item_->num_outputs == 0) {
GELOGD("Node[%s] type[%s] has no output, no need set mem_copy task.",
node_name_.c_str(), node_item_->node_type.c_str());
return SUCCESS;
}

GELOGD("Start to set memcpy task for node[%s].", node_name_.c_str());
const domi::KernelDef &kernel_def = task_def.kernel();
auto &memcpy_args = kernel_def.args();
uint32_t memcpy_args_size = kernel_def.args_size();
GE_IF_BOOL_EXEC(memcpy_args.size() != memcpy_args_size,
REPORT_INNER_ERROR("E19999", "MemCopy task def args.size=%zu, but args_size=%u not equal.",
memcpy_args.size(), memcpy_args_size);
GELOGE(FAILED, "[Check][Size]MemCopy task def args.size=%zu, but args_size=%u not equal.",
memcpy_args.size(), memcpy_args_size);
return FAILED;);
GE_IF_BOOL_EXEC(memcpy_args_size < sizeof(aicpu::AicpuParamHead),
REPORT_INNER_ERROR("E19999",
"Task def args_size=%u is less than aicpu param head len=%zu.",
memcpy_args_size, sizeof(aicpu::AicpuParamHead));
GELOGE(FAILED,
"[Check][Size] Task def args_size=%u is less than aicpu param head len=%zu.",
memcpy_args_size, sizeof(aicpu::AicpuParamHead));
return FAILED;);

std::unique_ptr<uint8_t[]> memcpy_task;
memcpy_task.reset(new(std::nothrow) uint8_t[memcpy_args_size]());
GE_IF_BOOL_EXEC(memcpy_task == nullptr,
REPORT_INNER_ERROR("E19999", "new memory failed for Node[MemCopy], task_size[%u].",
memcpy_args_size);
GELOGE(FAILED, "[Malloc][Memory] failed for Node[MemCopy], task_size[%u].",
memcpy_args_size);
return FAILED;);

errno_t sec_ret = memcpy_s(memcpy_task.get(), memcpy_args_size, memcpy_args.c_str(), memcpy_args.size());
GE_IF_BOOL_EXEC(sec_ret != EOK,
REPORT_INNER_ERROR("E19999",
"memcpy_s argc_ failed for Node[MemCopy], ret: %d", sec_ret);
GELOGE(INTERNAL_ERROR,
"[Update][args] failed for Node[MemCopy], ret: %d", sec_ret);
return sec_ret;);
GE_CHK_RT_RET(rtMemcpy(copy_task_args_buf_->GetData(), sizeof(aicpu::AicpuParamHead),
&memcpy_task, sizeof(aicpu::AicpuParamHead), RT_MEMCPY_HOST_TO_DEVICE));
GELOGD("Set memcpy task for node[%s] successfully.", node_name_.c_str());
return SUCCESS;
}

Status AicpuNodeTask::UpdateIoAddr(TaskContext &context) { Status AicpuNodeTask::UpdateIoAddr(TaskContext &context) {
vector<uint64_t> io_addrs; vector<uint64_t> io_addrs;
io_addrs.reserve(node_item_->num_inputs + node_item_->num_outputs); io_addrs.reserve(node_item_->num_inputs + node_item_->num_outputs);
@@ -829,15 +871,19 @@ Status AicpuNodeTask::LaunchTask(TaskContext &context) {
return SUCCESS; return SUCCESS;
} }


Status AicpuNodeTask::TaskCallback(TaskContext &context) {
Status AicpuNodeTaskBase::TaskCallback(TaskContext &context) {
GELOGD("Node[%s] task callback start, is_dynamic = %s, unknown_type=%d.", GELOGD("Node[%s] task callback start, is_dynamic = %s, unknown_type=%d.",
node_name_.c_str(), node_item_->is_dynamic ? "true" : "false", unknown_type_); node_name_.c_str(), node_item_->is_dynamic ? "true" : "false", unknown_type_);
Status callback_ret = SUCCESS; Status callback_ret = SUCCESS;


// check need update shape, call update shape. // check need update shape, call update shape.
if (node_item_->is_dynamic && unknown_type_ == DEPEND_SHAPE_RANGE) {
// check result
callback_ret = UpdateOutputShapeFromExtInfo(context);
if (node_item_->is_dynamic) {
if (unknown_type_ == DEPEND_SHAPE_RANGE) {
// check result
callback_ret = UpdateOutputShapeFromExtInfo(context);
} else if (unknown_type_ == DEPEND_COMPUTE) {
callback_ret = UpdateShapeAndDataByResultSummary(context);
}
} else { } else {
GELOGD("Node[%s] unknown shape type is %d no need update output shape.", GELOGD("Node[%s] unknown shape type is %d no need update output shape.",
node_name_.c_str(), unknown_type_); node_name_.c_str(), unknown_type_);
@@ -854,6 +900,31 @@ Status AiCpuNodeExecutor::PrepareTask(NodeTask &task, TaskContext &context) cons
return status; return status;
} }


Status AicpuNodeTask::UpdateShapeAndDataByResultSummary(TaskContext &context) {
GELOGD("Node[%s] update shape and data by result summary begin.", node_name_.c_str());

std::vector<std::unique_ptr<TensorBuffer>> out_shape_hbm;
GE_CHK_STATUS_RET(ReadResultSummaryAndPrepareMemory(context, out_shape_hbm),
"[Invoke][ReadResultSummaryAndPrepareMemory] failed for Node[%s].",
node_name_.c_str());

RECORD_CALLBACK_EVENT(context.GetExecutionContext(), node_name_.c_str(),
"[ReadResultSummaryAndPrepareMemory] End");

GE_CHK_STATUS_RET(CopyDataToHbm(context, out_shape_hbm),
"[Invoke][CopyDataToHbm] failed for Node[%s] copy data to output.",
node_name_.c_str());

RECORD_CALLBACK_EVENT(context.GetExecutionContext(), node_name_.c_str(), "[CopyDataToHbm] End");

GE_CHK_STATUS_RET(UpdateShapeByHbmBuffer(context, out_shape_hbm),
"[Update][ShapeByHbmBuffer] failed for Node[%s].",
node_name_.c_str());

GELOGD("Node[%s] update shape and data by result summary end.", node_name_.c_str());
return SUCCESS;
}

Status AiCpuNodeExecutor::LoadTask(const HybridModel &model, Status AiCpuNodeExecutor::LoadTask(const HybridModel &model,
const NodePtr &node, const NodePtr &node,
std::shared_ptr<NodeTask> &task) const { std::shared_ptr<NodeTask> &task) const {


+ 27
- 25
ge/hybrid/node_executor/aicpu/aicpu_node_executor.h View File

@@ -51,11 +51,17 @@ class AicpuNodeTaskBase : public NodeTask {


virtual Status UpdateOutputShapeFromExtInfo(TaskContext &task_context); virtual Status UpdateOutputShapeFromExtInfo(TaskContext &task_context);


virtual Status UpdateShapeAndDataByResultSummary(TaskContext &context) = 0;

Status UpdateShapeToOutputDesc(TaskContext &task_context, const GeShape &shape_new, int32_t output_index); Status UpdateShapeToOutputDesc(TaskContext &task_context, const GeShape &shape_new, int32_t output_index);


virtual Status LaunchTask(TaskContext &context) = 0; virtual Status LaunchTask(TaskContext &context) = 0;
Status InitForDependComputeTask(bool is_aicpu_kernel);


virtual Status TaskCallback(TaskContext &context) = 0;
Status TaskCallback(TaskContext &context);
virtual Status SetMemCopyTask(const domi::TaskDef &task_def) = 0;


virtual Status UpdateIoAddr(TaskContext &context) = 0; virtual Status UpdateIoAddr(TaskContext &context) = 0;


@@ -78,6 +84,20 @@ class AicpuNodeTaskBase : public NodeTask {


// ext info addr, device mem // ext info addr, device mem
std::unique_ptr<TensorBuffer> ext_info_addr_dev_; std::unique_ptr<TensorBuffer> ext_info_addr_dev_;

// just used for depend DEPEND_COMPUTE op
std::unique_ptr<TensorBuffer> copy_task_args_buf_;

std::vector<std::unique_ptr<TensorBuffer>> output_summary_;
std::vector<aicpu::FWKAdapter::ResultSummary> output_summary_host_;

std::unique_ptr<TensorBuffer> copy_ioaddr_dev_;

std::unique_ptr<TensorBuffer> copy_input_release_flag_dev_;
std::unique_ptr<TensorBuffer> copy_input_data_size_dev_;
std::unique_ptr<TensorBuffer> copy_input_src_dev_;
std::unique_ptr<TensorBuffer> copy_input_dst_dev_;
bool need_sync_ = false;
}; };


class AicpuTfNodeTask : public AicpuNodeTaskBase { class AicpuTfNodeTask : public AicpuNodeTaskBase {
@@ -93,17 +113,12 @@ class AicpuTfNodeTask : public AicpuNodeTaskBase {


Status LaunchTask(TaskContext &context) override; Status LaunchTask(TaskContext &context) override;


Status TaskCallback(TaskContext &context) override;

Status UpdateIoAddr(TaskContext &context) override; Status UpdateIoAddr(TaskContext &context) override;
Status UpdateShapeAndDataByResultSummary(TaskContext &context) override;


Status SetMemCopyTask(const domi::TaskDef &task_def) override;
private: private:
Status SetMemCopyTask(const domi::TaskDef &task_def);

Status InitForDependComputeTask();

Status UpdateShapeAndDataByResultSummary(TaskContext &context);

/// ///
/// read result summary and prepare copy task memory. /// read result summary and prepare copy task memory.
/// @param context task context /// @param context task context
@@ -132,20 +147,6 @@ class AicpuTfNodeTask : public AicpuNodeTaskBase {
// input and output addr, device mem // input and output addr, device mem
std::unique_ptr<TensorBuffer> input_output_addr_; std::unique_ptr<TensorBuffer> input_output_addr_;


// just used for depend DEPEND_COMPUTE op
std::unique_ptr<TensorBuffer> copy_task_args_buf_;

std::vector<std::unique_ptr<TensorBuffer>> output_summary_;
std::vector<aicpu::FWKAdapter::ResultSummary> output_summary_host_;

std::unique_ptr<TensorBuffer> copy_ioaddr_dev_;

std::unique_ptr<TensorBuffer> copy_input_release_flag_dev_;
std::unique_ptr<TensorBuffer> copy_input_data_size_dev_;
std::unique_ptr<TensorBuffer> copy_input_src_dev_;
std::unique_ptr<TensorBuffer> copy_input_dst_dev_;
bool need_sync_ = false;

std::unique_ptr<TensorBuffer> copy_workspace_buf_; std::unique_ptr<TensorBuffer> copy_workspace_buf_;
}; };


@@ -162,10 +163,11 @@ class AicpuNodeTask : public AicpuNodeTaskBase {


Status LaunchTask(TaskContext &context) override; Status LaunchTask(TaskContext &context) override;


Status TaskCallback(TaskContext &context) override;

Status UpdateIoAddr(TaskContext &context) override; Status UpdateIoAddr(TaskContext &context) override;
Status UpdateShapeAndDataByResultSummary(TaskContext &context) override;


Status SetMemCopyTask(const domi::TaskDef &task_def) override;
protected: protected:
// host mem // host mem
std::unique_ptr<uint8_t[]> args_; std::unique_ptr<uint8_t[]> args_;


Loading…
Cancel
Save