You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

davinci_model_unittest.cc 34 kB

4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
4 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877
  1. /**
  2. * Copyright 2019-2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include <gtest/gtest.h>
  17. #define private public
  18. #define protected public
  19. #include "graph/utils/graph_utils.h"
  20. #include "common/profiling/profiling_manager.h"
  21. #include "graph/load/model_manager/davinci_model.h"
  22. using namespace std;
  23. namespace ge {
  24. extern OpDescPtr CreateOpDesc(string name, string type);
  25. class UtestDavinciModel : public testing::Test {
  26. protected:
  27. void SetUp() {}
  28. void TearDown() {}
  29. };
  30. int32_t MsprofReport(uint32_t moduleId, uint32_t type, void *data, uint32_t len) {
  31. return 0;
  32. }
  33. TEST_F(UtestDavinciModel, init_success) {
  34. DavinciModel model(0, nullptr);
  35. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  36. ProfilingManager::Instance().is_load_profiling_ = true;
  37. GeModelPtr ge_model = make_shared<GeModel>();
  38. ge_model->SetGraph(GraphUtils::CreateGraphFromComputeGraph(graph));
  39. AttrUtils::SetInt(ge_model, ATTR_MODEL_MEMORY_SIZE, 10240);
  40. AttrUtils::SetInt(ge_model, ATTR_MODEL_STREAM_NUM, 1);
  41. shared_ptr<domi::ModelTaskDef> model_task_def = make_shared<domi::ModelTaskDef>();
  42. ge_model->SetModelTaskDef(model_task_def);
  43. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  44. TensorUtils::SetSize(tensor, 512);
  45. {
  46. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  47. op_desc->AddInputDesc(tensor);
  48. op_desc->AddOutputDesc(tensor);
  49. op_desc->SetInputOffset({1024});
  50. op_desc->SetOutputOffset({1024});
  51. NodePtr node = graph->AddNode(op_desc); // op_index = 0
  52. }
  53. {
  54. OpDescPtr op_desc = CreateOpDesc("square", "Square");
  55. op_desc->AddInputDesc(tensor);
  56. op_desc->AddOutputDesc(tensor);
  57. op_desc->SetInputOffset({1024});
  58. op_desc->SetOutputOffset({1024});
  59. NodePtr node = graph->AddNode(op_desc); // op_index = 1
  60. domi::TaskDef *task_def = model_task_def->add_task();
  61. task_def->set_stream_id(0);
  62. task_def->set_type(RT_MODEL_TASK_KERNEL);
  63. domi::KernelDef *kernel_def = task_def->mutable_kernel();
  64. kernel_def->set_stub_func("stub_func");
  65. kernel_def->set_args_size(64);
  66. string args(64, '1');
  67. kernel_def->set_args(args.data(), 64);
  68. domi::KernelContext *context = kernel_def->mutable_context();
  69. context->set_op_index(op_desc->GetId());
  70. context->set_kernel_type(2); // ccKernelType::TE
  71. uint16_t args_offset[9] = {0};
  72. context->set_args_offset(args_offset, 9 * sizeof(uint16_t));
  73. }
  74. {
  75. OpDescPtr op_desc = CreateOpDesc("memcpy", MEMCPYASYNC);
  76. op_desc->AddInputDesc(tensor);
  77. op_desc->AddOutputDesc(tensor);
  78. op_desc->SetInputOffset({1024});
  79. op_desc->SetOutputOffset({5120});
  80. NodePtr node = graph->AddNode(op_desc); // op_index = 2
  81. domi::TaskDef *task_def = model_task_def->add_task();
  82. task_def->set_stream_id(0);
  83. task_def->set_type(RT_MODEL_TASK_MEMCPY_ASYNC);
  84. domi::MemcpyAsyncDef *memcpy_async = task_def->mutable_memcpy_async();
  85. memcpy_async->set_src(1024);
  86. memcpy_async->set_dst(5120);
  87. memcpy_async->set_dst_max(512);
  88. memcpy_async->set_count(1);
  89. memcpy_async->set_kind(RT_MEMCPY_DEVICE_TO_DEVICE);
  90. memcpy_async->set_op_index(op_desc->GetId());
  91. }
  92. {
  93. OpDescPtr op_desc = CreateOpDesc("output", NETOUTPUT);
  94. op_desc->AddInputDesc(tensor);
  95. op_desc->SetInputOffset({5120});
  96. op_desc->SetSrcName( { "memcpy" } );
  97. op_desc->SetSrcIndex( { 0 } );
  98. NodePtr node = graph->AddNode(op_desc); // op_index = 3
  99. }
  100. EXPECT_EQ(model.Assign(ge_model), SUCCESS);
  101. EXPECT_EQ(model.Init(), SUCCESS);
  102. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  103. EXPECT_EQ(model.output_addrs_list_.size(), 1);
  104. EXPECT_EQ(model.task_list_.size(), 2);
  105. OutputData output_data;
  106. vector<OutputTensorInfo> outputs;
  107. EXPECT_EQ(model.GenOutputTensorInfo(&output_data, outputs), SUCCESS);
  108. EXPECT_EQ(output_data.blobs.size(), 1);
  109. EXPECT_EQ(outputs.size(), 1);
  110. ProfilingManager::Instance().is_load_profiling_ = false;
  111. }
  112. TEST_F(UtestDavinciModel, init_data_op) {
  113. DavinciModel model(0, nullptr);
  114. model.ge_model_ = make_shared<GeModel>();
  115. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  116. model.runtime_param_.mem_size = 5120000;
  117. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  118. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  119. TensorUtils::SetSize(tensor, 512);
  120. OpDescPtr op_input = CreateOpDesc("data", DATA);
  121. op_input->AddInputDesc(tensor);
  122. op_input->AddOutputDesc(tensor);
  123. op_input->SetInputOffset({1024});
  124. op_input->SetOutputOffset({1024});
  125. NodePtr node_input = graph->AddNode(op_input);
  126. OpDescPtr op_output = CreateOpDesc("output", NETOUTPUT);
  127. op_output->AddInputDesc(tensor);
  128. op_output->SetInputOffset({1024});
  129. op_output->SetSrcName( { "data" } );
  130. op_output->SetSrcIndex( { 0 } );
  131. NodePtr node_output = graph->AddNode(op_output);
  132. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  133. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  134. EXPECT_EQ(model.output_addrs_list_.size(), 1);
  135. EXPECT_EQ(model.op_list_.size(), 2);
  136. }
  137. TEST_F(UtestDavinciModel, init_data_op_subgraph) {
  138. DavinciModel model(0, nullptr);
  139. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  140. model.runtime_param_.mem_size = 5120000;
  141. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  142. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  143. TensorUtils::SetSize(tensor, 512);
  144. OpDescPtr op_input = CreateOpDesc("data", DATA);
  145. op_input->AddInputDesc(tensor);
  146. op_input->AddOutputDesc(tensor);
  147. op_input->SetInputOffset({1024});
  148. op_input->SetOutputOffset({1024});
  149. NodePtr node = graph->AddNode(op_input);
  150. uint32_t data_op_index = 0;
  151. map<uint32_t, OpDescPtr> data_by_index;
  152. set<const void *> input_outside_addrs;
  153. EXPECT_EQ(model.InitDataOp(nullptr, node, data_op_index, data_by_index, input_outside_addrs), SUCCESS);
  154. EXPECT_EQ(model.input_addrs_list_.size(), 0);
  155. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  156. EXPECT_EQ(data_op_index, 0);
  157. EXPECT_TRUE(data_by_index.empty());
  158. }
  159. TEST_F(UtestDavinciModel, init_netoutput_op_subgraph) {
  160. DavinciModel model(0, nullptr);
  161. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  162. model.runtime_param_.mem_size = 5120000;
  163. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  164. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  165. TensorUtils::SetSize(tensor, 512);
  166. OpDescPtr op_output = CreateOpDesc("output", NETOUTPUT);
  167. op_output->AddInputDesc(tensor);
  168. op_output->SetInputOffset({1024});
  169. op_output->SetSrcName( { "data" } );
  170. op_output->SetSrcIndex( { 0 } );
  171. NodePtr node = graph->AddNode(op_output);
  172. std::vector<OpDescPtr> output_op_list;
  173. set<const void *> output_outside_addrs;
  174. EXPECT_EQ(model.InitNetOutput(nullptr, node, output_op_list, output_outside_addrs), SUCCESS);
  175. EXPECT_EQ(model.input_addrs_list_.size(), 0);
  176. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  177. EXPECT_TRUE(output_op_list.empty());
  178. }
  179. TEST_F(UtestDavinciModel, init_unknown) {
  180. DavinciModel model(0, nullptr);
  181. model.SetKnownNode(true);
  182. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  183. GeModelPtr ge_model = make_shared<GeModel>();
  184. ge_model->SetGraph(GraphUtils::CreateGraphFromComputeGraph(graph));
  185. AttrUtils::SetInt(ge_model, ATTR_MODEL_MEMORY_SIZE, 5120000);
  186. AttrUtils::SetInt(ge_model, ATTR_MODEL_STREAM_NUM, 1);
  187. shared_ptr<domi::ModelTaskDef> model_task_def = make_shared<domi::ModelTaskDef>();
  188. ge_model->SetModelTaskDef(model_task_def);
  189. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  190. TensorUtils::SetSize(tensor, 512);
  191. OpDescPtr op_input = CreateOpDesc("data", DATA);
  192. op_input->AddInputDesc(tensor);
  193. op_input->AddOutputDesc(tensor);
  194. op_input->SetInputOffset({1024});
  195. op_input->SetOutputOffset({1024});
  196. NodePtr node_input = graph->AddNode(op_input); // op_index = 0
  197. OpDescPtr op_kernel = CreateOpDesc("square", "Square");
  198. op_kernel->AddInputDesc(tensor);
  199. op_kernel->AddOutputDesc(tensor);
  200. op_kernel->SetInputOffset({1024});
  201. op_kernel->SetOutputOffset({1024});
  202. NodePtr node_kernel = graph->AddNode(op_kernel); // op_index = 1
  203. OpDescPtr op_memcpy = CreateOpDesc("memcpy", MEMCPYASYNC);
  204. op_memcpy->AddInputDesc(tensor);
  205. op_memcpy->AddOutputDesc(tensor);
  206. op_memcpy->SetInputOffset({1024});
  207. op_memcpy->SetOutputOffset({5120});
  208. NodePtr node_memcpy = graph->AddNode(op_memcpy); // op_index = 2
  209. OpDescPtr op_output = CreateOpDesc("output", NETOUTPUT);
  210. op_output->AddInputDesc(tensor);
  211. op_output->SetInputOffset({5120});
  212. op_output->SetSrcName( { "memcpy" } );
  213. op_output->SetSrcIndex( { 0 } );
  214. NodePtr node_output = graph->AddNode(op_output); // op_index = 3
  215. domi::TaskDef *task_def1 = model_task_def->add_task();
  216. task_def1->set_stream_id(0);
  217. task_def1->set_type(RT_MODEL_TASK_KERNEL);
  218. domi::KernelDef *kernel_def = task_def1->mutable_kernel();
  219. kernel_def->set_stub_func("stub_func");
  220. kernel_def->set_args_size(64);
  221. string args(64, '1');
  222. kernel_def->set_args(args.data(), 64);
  223. domi::KernelContext *context = kernel_def->mutable_context();
  224. context->set_op_index(1);
  225. context->set_kernel_type(2); // ccKernelType::TE
  226. uint16_t args_offset[9] = {0};
  227. context->set_args_offset(args_offset, 9 * sizeof(uint16_t));
  228. domi::TaskDef *task_def2 = model_task_def->add_task();
  229. task_def2->set_stream_id(0);
  230. task_def2->set_type(RT_MODEL_TASK_MEMCPY_ASYNC);
  231. domi::MemcpyAsyncDef *memcpy_async = task_def2->mutable_memcpy_async();
  232. memcpy_async->set_src(1024);
  233. memcpy_async->set_dst(5120);
  234. memcpy_async->set_dst_max(512);
  235. memcpy_async->set_count(1);
  236. memcpy_async->set_kind(RT_MEMCPY_DEVICE_TO_DEVICE);
  237. memcpy_async->set_op_index(2);
  238. EXPECT_EQ(model.Assign(ge_model), SUCCESS);
  239. ProfilingManager::Instance().is_load_profiling_ = true;
  240. EXPECT_EQ(model.Init(), SUCCESS);
  241. ProfilingManager::Instance().is_load_profiling_ = false;
  242. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  243. EXPECT_EQ(model.output_addrs_list_.size(), 1);
  244. EXPECT_EQ(model.task_list_.size(), 2);
  245. EXPECT_EQ(model.task_list_[0]->UpdateArgs(), SUCCESS);
  246. EXPECT_EQ(model.task_list_[1]->UpdateArgs(), SUCCESS);
  247. vector<string> out_shape_info;
  248. model.GetModelAttr(out_shape_info);
  249. vector<InputOutputDescInfo> input_descs;
  250. vector<InputOutputDescInfo> output_descs;
  251. EXPECT_EQ(model.GetInputOutputDescInfo(input_descs, output_descs), SUCCESS);
  252. int32_t virtual_addr = 0;
  253. const vector<void *> inputs = { &virtual_addr };
  254. const vector<void *> outputs = { &virtual_addr };
  255. EXPECT_EQ(model.UpdateKnownNodeArgs(inputs, outputs), SUCCESS);
  256. }
  257. TEST_F(UtestDavinciModel, Init_variable_op) {
  258. DavinciModel model(0, nullptr);
  259. model.ge_model_ = make_shared<GeModel>();
  260. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  261. model.runtime_param_.mem_size = 5120000;
  262. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  263. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  264. TensorUtils::SetSize(tensor, 512);
  265. OpDescPtr var1 = CreateOpDesc("var1", VARIABLE);
  266. var1->AddInputDesc(tensor);
  267. var1->AddOutputDesc(tensor);
  268. var1->SetInputOffset({1024});
  269. var1->SetOutputOffset({1024});
  270. AttrUtils::SetBool(var1, VAR_ATTR_VAR_IS_BROADCAST, true);
  271. graph->AddNode(var1);
  272. OpDescPtr var2 = CreateOpDesc(NODE_NAME_GLOBAL_STEP, VARIABLE);
  273. var2->AddInputDesc(tensor);
  274. var2->AddOutputDesc(tensor);
  275. var2->SetInputOffset({1024});
  276. var2->SetOutputOffset({1024});
  277. graph->AddNode(var2);
  278. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  279. EXPECT_EQ(model.ReturnNoOutput(1), PARAM_INVALID);
  280. EXPECT_EQ(model.SyncVarData(), SUCCESS);
  281. }
  282. TEST_F(UtestDavinciModel, InitRealSizeAndShapeInfo_succ1) {
  283. DavinciModel model(0, nullptr);
  284. model.ge_model_ = make_shared<GeModel>();
  285. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  286. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  287. OpDescPtr op_output = CreateOpDesc("output_ascend_mbatch_batch_1", NETOUTPUT);
  288. op_output->AddInputDesc(tensor);
  289. op_output->SetInputOffset({1024});
  290. NodePtr node_output = graph->AddNode(op_output);
  291. EXPECT_EQ(model.InitRealSizeAndShapeInfo(graph, node_output), SUCCESS);
  292. }
  293. TEST_F(UtestDavinciModel, InitRealSizeAndShapeInfo_succ2) {
  294. DavinciModel model(0, nullptr);
  295. ComputeGraphPtr graph = std::make_shared<ComputeGraph>("test_graph");
  296. OpDescPtr data1 = CreateOpDesc("data1", DATA);
  297. GeTensorDesc shape_desc(GeShape({4,3,224,224}), FORMAT_NCHW, DT_FLOAT);
  298. data1->AddInputDesc(shape_desc);
  299. data1->AddOutputDesc(shape_desc);
  300. NodePtr data1_node = graph->AddNode(data1);
  301. OpDescPtr case_node = CreateOpDesc("case1", CASE);
  302. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  303. case_node->AddInputDesc(tensor);
  304. case_node->AddOutputDesc(tensor);
  305. NodePtr case1_node = graph->AddNode(case_node);
  306. OpDescPtr output = CreateOpDesc("output1", NETOUTPUT);
  307. output->AddInputDesc(tensor);
  308. output->SetSrcName( { "case1" } );
  309. output->SetSrcIndex( { 0 } );
  310. NodePtr output_node = graph->AddNode(output);
  311. GraphUtils::AddEdge(data1_node->GetOutDataAnchor(0), case1_node->GetInDataAnchor(0));
  312. GraphUtils::AddEdge(case1_node->GetOutDataAnchor(0), output_node->GetInDataAnchor(0));
  313. (void)AttrUtils::SetStr(output_node->GetOpDesc(), ATTR_ALL_GEARS_INFO, "1;2;4;8");
  314. (void)AttrUtils::SetBool(case_node, ATTR_INSERT_BY_MBATCH, true);
  315. model.is_getnext_sink_dynamic_ = false;
  316. model.is_online_infer_dynamic_ = true;
  317. auto ret = model.InitRealSizeAndShapeInfo(graph, output_node);
  318. // GetGearAndRealOutShapeInfo without ATTR_NAME_DYNAMIC_OUTPUT_DIMS
  319. EXPECT_EQ(ret, SUCCESS);
  320. vector<string> dynamic_output_dims = {"0,0,1,1,0,2,2,0,4,3,0,8"};
  321. (void)AttrUtils::SetListStr(output_node->GetOpDesc(), ATTR_NAME_DYNAMIC_OUTPUT_DIMS, dynamic_output_dims);
  322. ret = model.InitRealSizeAndShapeInfo(graph, output_node);
  323. EXPECT_EQ(ret, SUCCESS);
  324. }
  325. TEST_F(UtestDavinciModel, InitRealSizeAndShapeInfo_succ3) {
  326. DavinciModel model(0, nullptr);
  327. ComputeGraphPtr graph = std::make_shared<ComputeGraph>("test_graph");
  328. OpDescPtr data1 = CreateOpDesc("data1", DATA);
  329. GeTensorDesc shape_desc(GeShape({4,3,224,224}), FORMAT_NCHW, DT_FLOAT);
  330. data1->AddInputDesc(shape_desc);
  331. data1->AddOutputDesc(shape_desc);
  332. NodePtr data1_node = graph->AddNode(data1);
  333. OpDescPtr shape_node = CreateOpDesc("ascend_mbatch_get_dynamic_dims_node", GETDYNAMICDIMS);
  334. GeTensorDesc in_tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  335. GeTensorDesc out_tensor(GeShape({4,3}), FORMAT_NCHW, DT_FLOAT);
  336. shape_node->AddInputDesc(in_tensor);
  337. shape_node->AddOutputDesc(out_tensor);
  338. NodePtr get_dynamic_dims_node = graph->AddNode(shape_node);
  339. OpDescPtr output = CreateOpDesc("output1", NETOUTPUT);
  340. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  341. output->AddInputDesc(tensor);
  342. output->SetSrcName( { "data1", "ascend_mbatch_get_dynamic_dims_node" } );
  343. output->SetSrcIndex( { 0, 1 } );
  344. NodePtr output_node = graph->AddNode(output);
  345. GraphUtils::AddEdge(data1_node->GetOutDataAnchor(0), output_node->GetInDataAnchor(0));
  346. GraphUtils::AddEdge(get_dynamic_dims_node->GetOutDataAnchor(0), output_node->GetInDataAnchor(1));
  347. (void)AttrUtils::SetStr(output_node->GetOpDesc(), ATTR_ALL_GEARS_INFO, "1,3;;4,3;,3");
  348. model.is_getnext_sink_dynamic_ = true;
  349. model.is_online_infer_dynamic_ = false;
  350. auto ret = model.InitRealSizeAndShapeInfo(graph, output_node);
  351. EXPECT_EQ(ret, SUCCESS);
  352. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  353. model.runtime_param_.mem_size = 4;
  354. ret = model.InitRealSizeAndShapeInfo(graph, output_node);
  355. EXPECT_EQ(ret, SUCCESS);
  356. }
  357. TEST_F(UtestDavinciModel, init_data_aipp_info) {
  358. DavinciModel model(0, nullptr);
  359. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  360. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  361. model.runtime_param_.mem_size = 5120000;
  362. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  363. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  364. TensorUtils::SetSize(tensor, 512);
  365. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  366. op_desc->AddInputDesc(tensor);
  367. op_desc->AddOutputDesc(tensor);
  368. op_desc->SetInputOffset({1024});
  369. op_desc->SetOutputOffset({1024});
  370. NodePtr node = graph->AddNode(op_desc);
  371. GeAttrValue::NAMED_ATTRS aipp_attr;
  372. aipp_attr.SetAttr("aipp_mode", GeAttrValue::CreateFrom<GeAttrValue::INT>(domi::AippOpParams::dynamic));
  373. aipp_attr.SetAttr("related_input_rank", GeAttrValue::CreateFrom<GeAttrValue::INT>(0));
  374. aipp_attr.SetAttr("max_src_image_size", GeAttrValue::CreateFrom<GeAttrValue::INT>(2048));
  375. aipp_attr.SetAttr("support_rotation", GeAttrValue::CreateFrom<GeAttrValue::INT>(1));
  376. EXPECT_TRUE(AttrUtils::SetNamedAttrs(op_desc, ATTR_NAME_AIPP, aipp_attr));
  377. AippConfigInfo aipp_info;
  378. EXPECT_EQ(model.GetAippInfo(0, aipp_info), ACL_ERROR_GE_AIPP_NOT_EXIST);
  379. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  380. EXPECT_EQ(model.GetAippInfo(0, aipp_info), SUCCESS);
  381. EXPECT_EQ(aipp_info.aipp_mode, domi::AippOpParams::dynamic);
  382. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  383. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  384. EXPECT_EQ(model.op_list_.size(), 1);
  385. }
  386. TEST_F(UtestDavinciModel, init_data_aipp_static) {
  387. DavinciModel model(0, nullptr);
  388. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  389. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  390. model.runtime_param_.mem_size = 5120000;
  391. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  392. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  393. TensorUtils::SetSize(tensor, 512);
  394. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  395. op_desc->AddInputDesc(tensor);
  396. op_desc->AddOutputDesc(tensor);
  397. op_desc->SetInputOffset({1024});
  398. op_desc->SetOutputOffset({1024});
  399. NodePtr node = graph->AddNode(op_desc);
  400. AttrUtils::SetStr(op_desc, ATTR_DATA_RELATED_AIPP_MODE, "static_aipp");
  401. InputAippType aipp_type;
  402. size_t aipp_index = 0;
  403. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), PARAM_INVALID);
  404. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  405. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), SUCCESS);
  406. EXPECT_EQ(aipp_type, DATA_WITH_STATIC_AIPP);
  407. EXPECT_EQ(aipp_index, 0xFFFFFFFFu);
  408. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  409. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  410. EXPECT_EQ(model.op_list_.size(), 1);
  411. }
  412. TEST_F(UtestDavinciModel, init_data_aipp_dynamic) {
  413. DavinciModel model(0, nullptr);
  414. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  415. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  416. model.runtime_param_.mem_size = 5120000;
  417. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  418. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  419. TensorUtils::SetSize(tensor, 512);
  420. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  421. op_desc->AddInputDesc(tensor);
  422. op_desc->AddOutputDesc(tensor);
  423. op_desc->SetInputOffset({1024});
  424. op_desc->SetOutputOffset({1024});
  425. NodePtr node = graph->AddNode(op_desc); // op_index 0
  426. AttrUtils::SetStr(op_desc, ATTR_DATA_RELATED_AIPP_MODE, "dynamic_aipp");
  427. AttrUtils::SetStr(op_desc, ATTR_DATA_AIPP_DATA_NAME_MAP, "releated_aipp");
  428. InputAippType aipp_type;
  429. size_t aipp_index = 0;
  430. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), PARAM_INVALID);
  431. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  432. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), SUCCESS);
  433. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  434. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  435. EXPECT_EQ(model.op_list_.size(), 1);
  436. }
  437. TEST_F(UtestDavinciModel, init_data_aipp_releated) {
  438. DavinciModel model(0, nullptr);
  439. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  440. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  441. model.runtime_param_.mem_size = 5120000;
  442. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  443. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  444. TensorUtils::SetSize(tensor, 512);
  445. {
  446. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  447. op_desc->AddInputDesc(tensor);
  448. op_desc->AddOutputDesc(tensor);
  449. op_desc->SetInputOffset({1024});
  450. op_desc->SetOutputOffset({1024});
  451. NodePtr node = graph->AddNode(op_desc); // op_index 0
  452. AttrUtils::SetStr(op_desc, ATTR_DATA_RELATED_AIPP_MODE, "dynamic_aipp");
  453. AttrUtils::SetStr(op_desc, ATTR_DATA_AIPP_DATA_NAME_MAP, "releated_aipp");
  454. }
  455. {
  456. OpDescPtr op_desc = CreateOpDesc("releated_aipp", DATA);
  457. op_desc->AddInputDesc(tensor);
  458. op_desc->AddOutputDesc(tensor);
  459. op_desc->SetInputOffset({1024});
  460. op_desc->SetOutputOffset({1024});
  461. NodePtr node = graph->AddNode(op_desc); // op_index 1
  462. }
  463. InputAippType aipp_type;
  464. size_t aipp_index = 0;
  465. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), PARAM_INVALID);
  466. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  467. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), SUCCESS);
  468. EXPECT_EQ(aipp_type, DATA_WITH_DYNAMIC_AIPP);
  469. EXPECT_EQ(aipp_index, 1);
  470. EXPECT_EQ(model.input_addrs_list_.size(), 2);
  471. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  472. EXPECT_EQ(model.op_list_.size(), 2);
  473. }
  474. TEST_F(UtestDavinciModel, init_data_aipp_dynamic_conf) {
  475. DavinciModel model(0, nullptr);
  476. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  477. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  478. model.runtime_param_.mem_size = 5120000;
  479. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  480. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  481. TensorUtils::SetSize(tensor, 512);
  482. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  483. op_desc->AddInputDesc(tensor);
  484. op_desc->AddOutputDesc(tensor);
  485. op_desc->SetInputOffset({1024});
  486. op_desc->SetOutputOffset({1024});
  487. NodePtr node = graph->AddNode(op_desc); // op_index 0
  488. AttrUtils::SetStr(op_desc, ATTR_DATA_RELATED_AIPP_MODE, "dynamic_aipp_conf");
  489. InputAippType aipp_type;
  490. size_t aipp_index = 0;
  491. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), PARAM_INVALID);
  492. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  493. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), SUCCESS);
  494. EXPECT_EQ(aipp_type, DYNAMIC_AIPP_NODE);
  495. EXPECT_EQ(aipp_index, 0xFFFFFFFFU);
  496. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  497. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  498. EXPECT_EQ(model.op_list_.size(), 1);
  499. }
  500. TEST_F(UtestDavinciModel, init_data_aipp_dynamic_invalid) {
  501. DavinciModel model(0, nullptr);
  502. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  503. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  504. model.runtime_param_.mem_size = 5120000;
  505. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  506. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  507. TensorUtils::SetSize(tensor, 512);
  508. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  509. op_desc->AddInputDesc(tensor);
  510. op_desc->AddOutputDesc(tensor);
  511. op_desc->SetInputOffset({1024});
  512. op_desc->SetOutputOffset({1024});
  513. NodePtr node = graph->AddNode(op_desc); // op_index 0
  514. AttrUtils::SetStr(op_desc, ATTR_DATA_RELATED_AIPP_MODE, "dynamic_aipp_invalid");
  515. InputAippType aipp_type;
  516. size_t aipp_index = 0;
  517. EXPECT_EQ(model.GetAippType(0, aipp_type, aipp_index), PARAM_INVALID);
  518. EXPECT_EQ(model.InitNodes(graph), ACL_ERROR_GE_AIPP_MODE_INVALID);
  519. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  520. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  521. EXPECT_EQ(model.op_list_.size(), 1);
  522. }
  523. TEST_F(UtestDavinciModel, init_data_aipp_input_info_empty) {
  524. DavinciModel model(0, nullptr);
  525. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  526. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  527. model.runtime_param_.mem_size = 5120000;
  528. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  529. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  530. TensorUtils::SetSize(tensor, 512);
  531. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  532. op_desc->AddInputDesc(tensor);
  533. op_desc->AddOutputDesc(tensor);
  534. op_desc->SetInputOffset({1024});
  535. op_desc->SetOutputOffset({1024});
  536. NodePtr node = graph->AddNode(op_desc); // op_index 0
  537. vector<string> inputs = {};
  538. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_INPUTS, inputs);
  539. vector<string> outputs = {};
  540. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_OUTPUTS, outputs);
  541. OriginInputInfo orig_input_info;
  542. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), ACL_ERROR_GE_AIPP_NOT_EXIST);
  543. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  544. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), SUCCESS);
  545. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  546. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  547. EXPECT_EQ(model.op_list_.size(), 1);
  548. }
  549. TEST_F(UtestDavinciModel, init_data_aipp_input_info_normal) {
  550. DavinciModel model(0, nullptr);
  551. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  552. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  553. model.runtime_param_.mem_size = 5120000;
  554. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  555. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  556. TensorUtils::SetSize(tensor, 512);
  557. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  558. op_desc->AddInputDesc(tensor);
  559. op_desc->AddOutputDesc(tensor);
  560. op_desc->SetInputOffset({1024});
  561. op_desc->SetOutputOffset({1024});
  562. NodePtr node = graph->AddNode(op_desc); // op_index 0
  563. vector<string> inputs = { "NCHW:DT_FLOAT:TensorName:TensorSize:3:1,2,8" };
  564. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_INPUTS, inputs);
  565. vector<string> outputs = { "NCHW:DT_FLOAT:TensorName:TensorSize:3:1,2,8" };
  566. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_OUTPUTS, outputs);
  567. OriginInputInfo orig_input_info;
  568. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), ACL_ERROR_GE_AIPP_NOT_EXIST);
  569. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  570. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), SUCCESS);
  571. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  572. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  573. EXPECT_EQ(model.op_list_.size(), 1);
  574. }
  575. TEST_F(UtestDavinciModel, init_data_aipp_input_info_invalid) {
  576. DavinciModel model(0, nullptr);
  577. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  578. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  579. model.runtime_param_.mem_size = 5120000;
  580. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  581. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  582. TensorUtils::SetSize(tensor, 512);
  583. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  584. op_desc->AddInputDesc(tensor);
  585. op_desc->AddOutputDesc(tensor);
  586. op_desc->SetInputOffset({1024});
  587. op_desc->SetOutputOffset({1024});
  588. NodePtr node = graph->AddNode(op_desc); // op_index 0
  589. vector<string> inputs = { "NCHW:DT_FLOAT:TensorName" }; // Invalid
  590. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_INPUTS, inputs);
  591. vector<string> outputs = { "NCHW:DT_FLOAT:TensorName:TensorSize:3:1,2,8" };
  592. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_OUTPUTS, outputs);
  593. OriginInputInfo orig_input_info;
  594. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), ACL_ERROR_GE_AIPP_NOT_EXIST);
  595. EXPECT_EQ(model.InitNodes(graph), ACL_ERROR_GE_AIPP_MODE_INVALID);
  596. EXPECT_EQ(model.GetOrigInputInfo(0, orig_input_info), ACL_ERROR_GE_AIPP_NOT_EXIST);
  597. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  598. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  599. EXPECT_EQ(model.op_list_.size(), 1);
  600. }
  601. TEST_F(UtestDavinciModel, init_data_aipp_input_dims_normal) {
  602. DavinciModel model(0, nullptr);
  603. model.ge_model_ = make_shared<GeModel>(); // for CustAICPUKernelStore::GetCustAICPUKernelStore()
  604. model.runtime_param_.mem_base = (uint8_t *)0x08000000;
  605. model.runtime_param_.mem_size = 5120000;
  606. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  607. GeTensorDesc tensor(GeShape(), FORMAT_NCHW, DT_FLOAT);
  608. TensorUtils::SetSize(tensor, 512);
  609. OpDescPtr op_desc = CreateOpDesc("data", DATA);
  610. op_desc->AddInputDesc(tensor);
  611. op_desc->AddOutputDesc(tensor);
  612. op_desc->SetInputOffset({1024});
  613. op_desc->SetOutputOffset({1024});
  614. NodePtr node = graph->AddNode(op_desc); // op_index 0
  615. vector<string> inputs = { "NCHW:DT_FLOAT:TensorName:TensorSize:3:1,2,8" };
  616. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_INPUTS, inputs);
  617. vector<string> outputs = { "NCHW:DT_FLOAT:TensorName:TensorSize:3:1,2,8" };
  618. AttrUtils::SetListStr(op_desc, ATTR_NAME_AIPP_OUTPUTS, outputs);
  619. vector<InputOutputDims> input_dims;
  620. vector<InputOutputDims> output_dims;
  621. EXPECT_EQ(model.GetAllAippInputOutputDims(0, input_dims, output_dims), ACL_ERROR_GE_AIPP_NOT_EXIST);
  622. EXPECT_EQ(model.InitNodes(graph), SUCCESS);
  623. EXPECT_EQ(model.GetAllAippInputOutputDims(0, input_dims, output_dims), SUCCESS);
  624. EXPECT_EQ(input_dims.size(), 1);
  625. EXPECT_EQ(output_dims.size(), 1);
  626. EXPECT_EQ(model.input_addrs_list_.size(), 1);
  627. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  628. EXPECT_EQ(model.op_list_.size(), 1);
  629. }
  630. // test label_set_task Init
  631. TEST_F(UtestDavinciModel, label_task_success) {
  632. DavinciModel model(0, nullptr);
  633. ComputeGraphPtr graph = make_shared<ComputeGraph>("default");
  634. GeModelPtr ge_model = make_shared<GeModel>();
  635. ge_model->SetGraph(GraphUtils::CreateGraphFromComputeGraph(graph));
  636. AttrUtils::SetInt(ge_model, ATTR_MODEL_MEMORY_SIZE, 10240);
  637. AttrUtils::SetInt(ge_model, ATTR_MODEL_STREAM_NUM, 1);
  638. shared_ptr<domi::ModelTaskDef> model_task_def = make_shared<domi::ModelTaskDef>();
  639. ge_model->SetModelTaskDef(model_task_def);
  640. GeTensorDesc tensor(GeShape(), FORMAT_ND, DT_INT32);
  641. TensorUtils::SetSize(tensor, 64);
  642. {
  643. OpDescPtr op_desc = CreateOpDesc("label_switch", LABELSWITCHBYINDEX);
  644. op_desc->AddInputDesc(tensor);
  645. op_desc->SetInputOffset({1024});
  646. NodePtr node = graph->AddNode(op_desc); // op_index = 0
  647. EXPECT_TRUE(AttrUtils::SetListInt(op_desc, ATTR_NAME_LABEL_SWITCH_LIST, {0, 1}));
  648. domi::TaskDef *task_def1 = model_task_def->add_task();
  649. task_def1->set_stream_id(0);
  650. task_def1->set_type(RT_MODEL_TASK_STREAM_LABEL_SWITCH_BY_INDEX);
  651. domi::LabelSwitchByIndexDef *label_task_def = task_def1->mutable_label_switch_by_index();
  652. label_task_def->set_op_index(op_desc->GetId());
  653. label_task_def->set_label_max(2);
  654. }
  655. {
  656. OpDescPtr op_desc = CreateOpDesc("label_then", LABELSET);
  657. NodePtr node = graph->AddNode(op_desc); // op_index = 1
  658. EXPECT_TRUE(AttrUtils::SetInt(op_desc, ATTR_NAME_LABEL_SWITCH_INDEX, 1));
  659. domi::TaskDef *task_def1 = model_task_def->add_task();
  660. task_def1->set_stream_id(0);
  661. task_def1->set_type(RT_MODEL_TASK_LABEL_SET);
  662. domi::LabelSetDef *label_task_def = task_def1->mutable_label_set();
  663. label_task_def->set_op_index(op_desc->GetId());
  664. }
  665. {
  666. OpDescPtr op_desc = CreateOpDesc("label_goto", LABELGOTOEX);
  667. NodePtr node = graph->AddNode(op_desc); // op_index = 2
  668. EXPECT_TRUE(AttrUtils::SetInt(op_desc, ATTR_NAME_LABEL_SWITCH_INDEX, 2));
  669. domi::TaskDef *task_def2 = model_task_def->add_task();
  670. task_def2->set_stream_id(0);
  671. task_def2->set_type(RT_MODEL_TASK_STREAM_LABEL_GOTO);
  672. domi::LabelGotoExDef *label_task_def = task_def2->mutable_label_goto_ex();
  673. label_task_def->set_op_index(op_desc->GetId());
  674. }
  675. {
  676. OpDescPtr op_desc = CreateOpDesc("label_else", LABELSET);
  677. NodePtr node = graph->AddNode(op_desc); // op_index = 3
  678. EXPECT_TRUE(AttrUtils::SetInt(op_desc, ATTR_NAME_LABEL_SWITCH_INDEX, 0));
  679. domi::TaskDef *task_def1 = model_task_def->add_task();
  680. task_def1->set_stream_id(0);
  681. task_def1->set_type(RT_MODEL_TASK_LABEL_SET);
  682. domi::LabelSetDef *label_task_def = task_def1->mutable_label_set();
  683. label_task_def->set_op_index(op_desc->GetId());
  684. }
  685. {
  686. OpDescPtr op_desc = CreateOpDesc("label_leave", LABELSET);
  687. NodePtr node = graph->AddNode(op_desc); // op_index = 4
  688. EXPECT_TRUE(AttrUtils::SetInt(op_desc, ATTR_NAME_LABEL_SWITCH_INDEX, 2));
  689. domi::TaskDef *task_def1 = model_task_def->add_task();
  690. task_def1->set_stream_id(0);
  691. task_def1->set_type(RT_MODEL_TASK_LABEL_SET);
  692. domi::LabelSetDef *label_task_def = task_def1->mutable_label_set();
  693. label_task_def->set_op_index(op_desc->GetId());
  694. }
  695. EXPECT_TRUE(AttrUtils::SetInt(ge_model, ATTR_MODEL_LABEL_NUM, 3));
  696. EXPECT_EQ(model.Assign(ge_model), SUCCESS);
  697. EXPECT_EQ(model.Init(), SUCCESS);
  698. EXPECT_EQ(model.input_addrs_list_.size(), 0);
  699. EXPECT_EQ(model.output_addrs_list_.size(), 0);
  700. EXPECT_EQ(model.task_list_.size(), 5);
  701. }
  702. TEST_F(UtestDavinciModel, LoadWithQueue_fail_with_diff_args) {
  703. DavinciModel model(0, nullptr);
  704. model.ge_model_ = make_shared<GeModel>();
  705. model.input_queue_ids_.emplace_back(0);
  706. EXPECT_EQ(model.LoadWithQueue(), ACL_ERROR_GE_EXEC_MODEL_QUEUE_ID_INVALID);
  707. EXPECT_EQ(model.input_data_info_.size(), 0);
  708. ZeroCopyOffset zero_copy_offset;
  709. model.input_data_info_[0] = zero_copy_offset;
  710. model.output_queue_ids_.emplace_back(0);
  711. EXPECT_EQ(model.LoadWithQueue(), ACL_ERROR_GE_EXEC_MODEL_QUEUE_ID_INVALID);
  712. EXPECT_EQ(model.output_data_info_.size(), 0);
  713. model.output_data_info_[0] = zero_copy_offset;
  714. EXPECT_EQ(model.LoadWithQueue(), INTERNAL_ERROR);
  715. EXPECT_EQ(model.active_stream_list_.size(), 0);
  716. }
  717. TEST_F(UtestDavinciModel, Sink_model_profile) {
  718. ProfilingManager::Instance().prof_cb_.msprofReporterCallback = MsprofReport;
  719. ProfileInfo profile;
  720. profile.fusion_info.op_name = "relu";
  721. DavinciModel model(0, nullptr);
  722. model.profile_list_.emplace_back(profile);
  723. std::map<std::string, std::pair<uint32_t, uint32_t>> op_info;
  724. op_info["relu"] = std::pair<uint32_t, uint32_t>(1, 1);
  725. model.profiler_report_op_info_ = op_info;
  726. model.SinkModelProfile();
  727. }
  728. } // namespace ge

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示