You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

mem_assigner_unittest.cc 11 kB

5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251
  1. /**
  2. * Copyright 2019-2020 Huawei Technologies Co., Ltd
  3. *
  4. * Licensed under the Apache License, Version 2.0 (the "License");
  5. * you may not use this file except in compliance with the License.
  6. * You may obtain a copy of the License at
  7. *
  8. * http://www.apache.org/licenses/LICENSE-2.0
  9. *
  10. * Unless required by applicable law or agreed to in writing, software
  11. * distributed under the License is distributed on an "AS IS" BASIS,
  12. * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. * See the License for the specific language governing permissions and
  14. * limitations under the License.
  15. */
  16. #include <gtest/gtest.h>
  17. #include <memory>
  18. #include "graph/anchor.h"
  19. #include "graph/attr_value.h"
  20. #include "graph/debug/ge_attr_define.h"
  21. #include "graph/utils/graph_utils.h"
  22. #include "graph/utils/node_utils.h"
  23. #include "graph/utils/op_desc_utils.h"
  24. #include "graph/utils/tensor_utils.h"
  25. #include "omg/omg_inner_types.h"
  26. #include "../passes/graph_builder_utils.h"
  27. #define protected public
  28. #define private public
  29. #include "graph/build/memory/binary_block_mem_assigner.h"
  30. #include "graph/build/memory/graph_mem_assigner.h"
  31. #include "graph/build/memory/hybrid_mem_assigner.h"
  32. #include "graph/build/memory/max_block_mem_assigner.h"
  33. #undef protected
  34. #undef private
  35. using namespace std;
  36. using namespace testing;
  37. using namespace ge;
  38. using domi::GetContext;
  39. class UtestMemoryAssignerTest : public testing::Test {
  40. public:
  41. ge::OpDescPtr CreateOpWithWsSize(const string &name, int64_t wsByte, const string &type = "some") {
  42. ge::OpDescPtr op_def = make_shared<ge::OpDesc>(name, type);
  43. auto desc_temp_ptr = make_shared<ge::GeTensorDesc>();
  44. auto desc_temp = *desc_temp_ptr;
  45. TensorUtils::SetSize(desc_temp, 1024);
  46. op_def->AddInputDesc(desc_temp);
  47. op_def->AddOutputDesc(desc_temp);
  48. std::vector<int64_t> workspace_bytes;
  49. workspace_bytes.push_back(wsByte);
  50. op_def->SetWorkspaceBytes(workspace_bytes);
  51. return op_def;
  52. }
  53. ge::OpDescPtr CreateRefOpWithWsSize(const string &name, int64_t wsByte, const string &type = "some") {
  54. ge::OpDescPtr op_def = make_shared<ge::OpDesc>(name, type);
  55. auto desc_temp_ptr = make_shared<ge::GeTensorDesc>();
  56. auto desc_temp = *desc_temp_ptr;
  57. TensorUtils::SetSize(desc_temp, 1024);
  58. op_def->AddInputDesc(desc_temp);
  59. auto desc_output_ptr = make_shared<ge::GeTensorDesc>();
  60. auto desc_output = *desc_output_ptr;
  61. TensorUtils::SetSize(desc_output, 6500);
  62. ge::TensorUtils::SetReuseInput(desc_output, true);
  63. ge::TensorUtils::SetReuseInputIndex(desc_output, 0);
  64. op_def->AddOutputDesc(desc_output);
  65. std::vector<int64_t> workspace_bytes;
  66. workspace_bytes.push_back(wsByte);
  67. op_def->SetWorkspaceBytes(workspace_bytes);
  68. return op_def;
  69. }
  70. void MakeGraph(ge::ComputeGraphPtr &graph) {
  71. ge::OpDescPtr op_def_a = CreateOpWithWsSize("A", 6000);
  72. op_def_a->SetStreamId(0);
  73. ge::OpDescPtr op_def_b = CreateOpWithWsSize("B", 120000);
  74. op_def_b->SetStreamId(0);
  75. ge::OpDescPtr op_def_c = CreateOpWithWsSize("C", 16000);
  76. op_def_c->SetStreamId(1);
  77. ge::OpDescPtr op_def_d = CreateOpWithWsSize("D", 24000);
  78. op_def_d->SetStreamId(2);
  79. ge::OpDescPtr op_def_e = CreateOpWithWsSize("E", 24000);
  80. op_def_e->SetStreamId(3);
  81. ge::OpDescPtr op_def_f = CreateOpWithWsSize("F", 30000);
  82. op_def_f->SetStreamId(2);
  83. ge::OpDescPtr op_def_g = CreateOpWithWsSize("G", 32000);
  84. op_def_g->SetStreamId(3);
  85. ge::OpDescPtr op_def_h = CreateOpWithWsSize("H", 48000);
  86. op_def_h->SetStreamId(2);
  87. ge::OpDescPtr op_def_i = CreateOpWithWsSize("I", 60000);
  88. op_def_i->SetStreamId(2);
  89. ge::OpDescPtr op_def_j = CreateOpWithWsSize("J", 256000, NETOUTPUT);
  90. op_def_j->SetStreamId(3);
  91. // add node
  92. ge::NodePtr node_a = graph->AddNode(op_def_a);
  93. ge::NodePtr node_b = graph->AddNode(op_def_b);
  94. ge::NodePtr node_c = graph->AddNode(op_def_c);
  95. ge::NodePtr node_d = graph->AddNode(op_def_d);
  96. ge::NodePtr node_e = graph->AddNode(op_def_e);
  97. ge::NodePtr node_f = graph->AddNode(op_def_f);
  98. ge::NodePtr node_g = graph->AddNode(op_def_g);
  99. ge::NodePtr node_h = graph->AddNode(op_def_h);
  100. ge::NodePtr node_i = graph->AddNode(op_def_i);
  101. ge::NodePtr node_j = graph->AddNode(op_def_j);
  102. // add edge
  103. ge::GraphUtils::AddEdge(node_a->GetOutDataAnchor(0), node_b->GetInDataAnchor(0));
  104. ge::GraphUtils::AddEdge(node_a->GetOutDataAnchor(0), node_c->GetInDataAnchor(0));
  105. ge::GraphUtils::AddEdge(node_b->GetOutDataAnchor(0), node_d->GetInDataAnchor(0));
  106. ge::GraphUtils::AddEdge(node_b->GetOutDataAnchor(0), node_e->GetInDataAnchor(0));
  107. ge::GraphUtils::AddEdge(node_c->GetOutDataAnchor(0), node_g->GetInDataAnchor(0));
  108. ge::GraphUtils::AddEdge(node_d->GetOutDataAnchor(0), node_f->GetInDataAnchor(0));
  109. ge::GraphUtils::AddEdge(node_e->GetOutDataAnchor(0), node_g->GetInDataAnchor(1));
  110. ge::GraphUtils::AddEdge(node_f->GetOutDataAnchor(0), node_h->GetInDataAnchor(0));
  111. ge::GraphUtils::AddEdge(node_g->GetOutDataAnchor(0), node_j->GetInDataAnchor(0));
  112. ge::GraphUtils::AddEdge(node_h->GetOutDataAnchor(0), node_i->GetInDataAnchor(0));
  113. ge::GraphUtils::AddEdge(node_i->GetOutDataAnchor(0), node_j->GetInDataAnchor(1));
  114. GetContext().out_nodes_map["H"] = {0};
  115. GetContext().out_nodes_map["I"] = {0};
  116. GetContext().out_nodes_map["J"] = {0};
  117. graph->TopologicalSorting();
  118. }
  119. void MakeReuseGraph(ge::ComputeGraphPtr graph) {
  120. ge::OpDescPtr op_def_a = CreateOpWithWsSize("A", 6000);
  121. ge::OpDescPtr op_def_b = CreateOpWithWsSize("B", 120000);
  122. ge::OpDescPtr op_def_c = CreateRefOpWithWsSize("C", 120000);
  123. ge::OpDescPtr op_def_d = make_shared<ge::OpDesc>("D", "CONSTANT");
  124. ge::NodePtr node_a = graph->AddNode(op_def_a);
  125. ge::NodePtr node_b = graph->AddNode(op_def_b);
  126. ge::NodePtr node_c = graph->AddNode(op_def_c);
  127. ge::NodePtr node_d = graph->AddNode(op_def_d);
  128. ge::GraphUtils::AddEdge(node_a->GetOutDataAnchor(0), node_b->GetInDataAnchor(0));
  129. ge::GraphUtils::AddEdge(node_a->GetOutDataAnchor(0), node_c->GetInDataAnchor(0));
  130. ge::GraphUtils::AddEdge(node_a->GetOutDataAnchor(0), node_d->GetInDataAnchor(0));
  131. GetContext().out_nodes_map["B"] = {0};
  132. GetContext().out_nodes_map["C"] = {0};
  133. graph->TopologicalSorting();
  134. }
  135. ComputeGraphPtr MakeCascadeContinuousMemoryGraph() {
  136. ge::ut::GraphBuilder builder("graph");
  137. auto data = builder.AddNode("data", "Data", 1, 1);
  138. auto addn1 = builder.AddNode("addn1", "AddN", 1, 1);
  139. auto addn2 = builder.AddNode("addn2", "AddN", 1, 1);
  140. auto addn3 = builder.AddNode("addn3", "AddN", 1, 1);
  141. auto concat1 = builder.AddNode("concat1", "Concat", 2, 1);
  142. auto concat2 = builder.AddNode("concat2", "Concat", 2, 1);
  143. auto netoutput = builder.AddNode("netoutput", "NetOutput", 2, 0);
  144. ge::AttrUtils::SetBool(concat1->GetOpDesc(), ATTR_NAME_NOPADDING_CONTINUOUS_INPUT, true);
  145. ge::AttrUtils::SetBool(concat1->GetOpDesc(), ATTR_NAME_CONTINUOUS_INPUT_ALLOC, true);
  146. ge::AttrUtils::SetBool(concat1->GetOpDesc(), ATTR_NAME_OUTPUT_REUSE_INPUT, true);
  147. ge::AttrUtils::SetBool(concat2->GetOpDesc(), ATTR_NAME_NOPADDING_CONTINUOUS_INPUT, true);
  148. ge::AttrUtils::SetBool(concat2->GetOpDesc(), ATTR_NAME_CONTINUOUS_INPUT_ALLOC, true);
  149. ge::AttrUtils::SetBool(concat2->GetOpDesc(), ATTR_NAME_OUTPUT_REUSE_INPUT, true);
  150. addn1->GetOpDesc()->SetOutputOffset({100});
  151. addn2->GetOpDesc()->SetOutputOffset({200});
  152. concat1->GetOpDesc()->SetOutputOffset({100});
  153. addn3->GetOpDesc()->SetOutputOffset({700});
  154. concat2->GetOpDesc()->SetOutputOffset({500});
  155. ge::AttrUtils::SetListInt(addn1->GetOpDesc(), ATTR_NAME_OUTPUT_OFFSET_FOR_BUFFER_FUSION, {100});
  156. ge::AttrUtils::SetListInt(addn2->GetOpDesc(), ATTR_NAME_OUTPUT_OFFSET_FOR_BUFFER_FUSION, {100});
  157. ge::AttrUtils::SetListInt(addn3->GetOpDesc(), ATTR_NAME_OUTPUT_OFFSET_FOR_BUFFER_FUSION, {100});
  158. ge::AttrUtils::SetListInt(concat1->GetOpDesc(), ATTR_NAME_OUTPUT_OFFSET_FOR_BUFFER_FUSION, {200});
  159. ge::AttrUtils::SetListInt(concat2->GetOpDesc(), ATTR_NAME_OUTPUT_OFFSET_FOR_BUFFER_FUSION, {300});
  160. builder.AddDataEdge(data, 0, addn1, 0);
  161. builder.AddDataEdge(data, 0, addn2, 0);
  162. builder.AddDataEdge(addn1, 0, concat1, 0);
  163. builder.AddDataEdge(addn2, 0, concat1, 1);
  164. builder.AddDataEdge(concat1, 0, concat2, 0);
  165. builder.AddDataEdge(addn3, 0, concat2, 1);
  166. return builder.GetGraph();
  167. }
  168. protected:
  169. void SetUp() {}
  170. void TearDown() { GetContext().out_nodes_map.clear(); }
  171. };
  172. /*
  173. TEST_F(UtestMemoryAssignerTest, MemoryBlock_Resize_RealSizeList_is_empty) {
  174. ge::ComputeGraphPtr graph = make_shared<ge::ComputeGraph>("");
  175. ge::OpDescPtr op_def_a = CreateOpWithWsSize("A", 6000);
  176. ge::NodePtr node_a = graph->AddNode(op_def_a);
  177. MemoryBlock* memory_block = new MemoryBlock(0);
  178. memory_block->Init(1, kOutput, node_a, 0, 1);
  179. memory_block->real_size_list_.clear();
  180. memory_block->Resize();
  181. EXPECT_EQ(memory_block->Size(), 0);
  182. delete memory_block;
  183. }
  184. */
  185. namespace ge {
  186. class MockBlockMemAssigner : public BlockMemAssigner {
  187. public:
  188. explicit MockBlockMemAssigner(ge::ComputeGraphPtr compute_graph, const std::map<std::string, std::string> &anchor_to_symbol, const std::map<std::string, std::list<NodeIndexIO>> &symbol_to_anchors) : BlockMemAssigner(compute_graph, anchor_to_symbol, symbol_to_anchors) {};
  189. virtual ~MockBlockMemAssigner(){};
  190. Status GetMemoryRanges(std::vector<int64_t> &ranges) override { return FAILED; }
  191. };
  192. } // namespace ge
  193. // when check GetMemoryRanges return fail, Assign return fail
  194. TEST_F(UtestMemoryAssignerTest, Mock_block_mem_assigner_failed) {
  195. ge::ComputeGraphPtr graph = make_shared<ge::ComputeGraph>("");
  196. MakeGraph(graph);
  197. std::map<std::string, std::string> anchor_to_symbol;
  198. std::map<std::string, std::list<NodeIndexIO>> symbol_to_anchors;
  199. EXPECT_EQ(GraphUtils::GetRefMapping(graph, symbol_to_anchors, anchor_to_symbol), GRAPH_SUCCESS);
  200. MockBlockMemAssigner mock_assigner(graph, anchor_to_symbol, symbol_to_anchors);
  201. EXPECT_EQ(mock_assigner.Assign(), FAILED);
  202. }
  203. TEST_F(UtestMemoryAssignerTest, graph_memory_assign_continuous_input) {
  204. ge::ComputeGraphPtr graph = MakeCascadeContinuousMemoryGraph();
  205. auto addn1 = graph->FindNode("addn1");
  206. auto addn2 = graph->FindNode("addn2");
  207. EXPECT_EQ(addn1->GetOpDesc()->GetOutputOffset()[0], 100);
  208. EXPECT_EQ(addn2->GetOpDesc()->GetOutputOffset()[0], 200);
  209. GraphMemoryAssigner memoryAssigner(graph);
  210. MemoryOffset memory_offset(RT_MEMORY_HBM, 0);
  211. memoryAssigner.memory_offset_.emplace(RT_MEMORY_HBM, memory_offset);
  212. EXPECT_EQ(memoryAssigner.ReAssignContinuousMemory(false), GRAPH_SUCCESS);
  213. EXPECT_EQ(addn1->GetOpDesc()->GetOutputOffset()[0], 500);
  214. EXPECT_EQ(addn2->GetOpDesc()->GetOutputOffset()[0], 600);
  215. }

图引擎模块(GE)是MindSpore的一个子模块,其代码由C++实现,位于前端模块ME和底层硬件之间,起到承接作用。图引擎模块以ME下发的图作为输入,然后进行一系列的深度图优化操作,最后输出一张可以在底层硬件上高效运行的图。GE针对昇腾AI处理器的硬件结构特点,做了特定的优化工作,以此来充分发挥出昇腾AI处理器的强大算力。在进行模型训练/推理时,GE会被自动调用而用户并不感知。GE主要由GE API和GE Core两部分组成,详细的架构图如下所示