@@ -41,6 +41,7 @@ if (ENABLE_OPEN_SRC) | |||||
message(FATAL_ERROR "Running on a unsupported architecture: ${SYSTEM_TYPE}, build terminated") | message(FATAL_ERROR "Running on a unsupported architecture: ${SYSTEM_TYPE}, build terminated") | ||||
endif() | endif() | ||||
set(GE_LIB_PATH ${GE_LIB_PATH}/${GE_SYS_ARCH}) | set(GE_LIB_PATH ${GE_LIB_PATH}/${GE_SYS_ARCH}) | ||||
set(STATIC_ACL_LIB ${GE_LIB_PATH}) | |||||
find_module(slog libslog.so ${GE_LIB_PATH}) | find_module(slog libslog.so ${GE_LIB_PATH}) | ||||
find_module(mmpa libmmpa.so ${GE_LIB_PATH}) | find_module(mmpa libmmpa.so ${GE_LIB_PATH}) | ||||
find_module(msprof libmsprof.so ${GE_LIB_PATH}) | find_module(msprof libmsprof.so ${GE_LIB_PATH}) | ||||
@@ -53,7 +54,7 @@ if (ENABLE_OPEN_SRC) | |||||
find_module(ascend_hal_stub libascend_hal.so ${GE_LIB_PATH}) | find_module(ascend_hal_stub libascend_hal.so ${GE_LIB_PATH}) | ||||
find_module(error_manager_static liberror_manager.a ${GE_LIB_PATH}) | find_module(error_manager_static liberror_manager.a ${GE_LIB_PATH}) | ||||
find_module(msprofiler libmsprofiler.a ${GE_LIB_PATH}) | find_module(msprofiler libmsprofiler.a ${GE_LIB_PATH}) | ||||
find_module(ascendcl_static libascendcl.a ${GE_LIB_PATH}) | |||||
#find_module(ascendcl_static libascendcl.a ${GE_LIB_PATH}) | |||||
else() | else() | ||||
if(DEFINED ENV{ASCEND_CUSTOM_PATH}) | if(DEFINED ENV{ASCEND_CUSTOM_PATH}) | ||||
set(ASCEND_DIR $ENV{ASCEND_CUSTOM_PATH}) | set(ASCEND_DIR $ENV{ASCEND_CUSTOM_PATH}) | ||||
@@ -66,6 +67,7 @@ if (ENABLE_OPEN_SRC) | |||||
set(ASCEND_RUNTIME_DIR ${ASCEND_DIR}/fwkacllib/lib64) | set(ASCEND_RUNTIME_DIR ${ASCEND_DIR}/fwkacllib/lib64) | ||||
set(ASCEND_ATC_DIR ${ASCEND_DIR}/atc/lib64) | set(ASCEND_ATC_DIR ${ASCEND_DIR}/atc/lib64) | ||||
set(ASCEND_ACL_DIR ${ASCEND_DIR}/acllib/lib64) | set(ASCEND_ACL_DIR ${ASCEND_DIR}/acllib/lib64) | ||||
set(STATIC_ACL_LIB ${ASCEND_ACL_DIR}) | |||||
find_module(slog libslog.so ${ASCEND_ATC_DIR}) | find_module(slog libslog.so ${ASCEND_ATC_DIR}) | ||||
find_module(mmpa libmmpa.so ${ASCEND_ATC_DIR}) | find_module(mmpa libmmpa.so ${ASCEND_ATC_DIR}) | ||||
if(PLATFORM STREQUAL "train") | if(PLATFORM STREQUAL "train") | ||||
@@ -88,15 +90,19 @@ if (ENABLE_OPEN_SRC) | |||||
find_module(error_manager liberror_manager.so ${ASCEND_ATC_DIR}) | find_module(error_manager liberror_manager.so ${ASCEND_ATC_DIR}) | ||||
find_module(error_manager_static liberror_manager.a ${ASCEND_ACL_DIR}) | find_module(error_manager_static liberror_manager.a ${ASCEND_ACL_DIR}) | ||||
find_module(msprofiler libmsprofiler.a ${ASCEND_ACL_DIR}) | find_module(msprofiler libmsprofiler.a ${ASCEND_ACL_DIR}) | ||||
find_module(ascendcl_static libascendcl.a ${ASCEND_ACL_DIR}) | |||||
if(NOT PRODUCT STREQUAL "flr3") | |||||
#find_module(ascendcl_static libascendcl.a ${ASCEND_ACL_DIR}) | |||||
if(PRODUCT STREQUAL "flr3") | |||||
find_module(msprof libmsprof.so ${ASCEND_DRIVER_SHARE_DIR}) | |||||
elseif(PRODUCT STREQUAL "flr1") | |||||
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}/driver) | |||||
find_module(msprof libmsprof.so ${ASCEND_DRIVER_COMMON_DIR}) | |||||
elseif(PRODUCT STREQUAL "flr2") | |||||
# flr2 ascend_hal_stub limsprof ? | |||||
else() | |||||
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}) | find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}) | ||||
find_module(msprof libmsprof.so ${ASCEND_DRIVER_DIR}) | find_module(msprof libmsprof.so ${ASCEND_DRIVER_DIR}) | ||||
else() | |||||
find_module(msprof libmsprof.so ${ASCEND_DRIVER_SHARE_DIR}) | |||||
endif() | endif() | ||||
elseif(PLATFORM STREQUAL "all") | elseif(PLATFORM STREQUAL "all") | ||||
#mdc没有 dc在driver目录 | |||||
find_module(msprof libmsprof.so ${ASCEND_DRIVER_COMMON_DIR}) | find_module(msprof libmsprof.so ${ASCEND_DRIVER_COMMON_DIR}) | ||||
find_module(hccl libhccl.so ${ASCEND_RUNTIME_DIR}) | find_module(hccl libhccl.so ${ASCEND_RUNTIME_DIR}) | ||||
find_module(adump_server libadump_server.a ${ASCEND_ACL_DIR}) | find_module(adump_server libadump_server.a ${ASCEND_ACL_DIR}) | ||||
@@ -106,9 +112,8 @@ if (ENABLE_OPEN_SRC) | |||||
find_module(error_manager liberror_manager.so ${ASCEND_ATC_DIR}) | find_module(error_manager liberror_manager.so ${ASCEND_ATC_DIR}) | ||||
find_module(error_manager_static liberror_manager.a ${ASCEND_ACL_DIR}) | find_module(error_manager_static liberror_manager.a ${ASCEND_ACL_DIR}) | ||||
find_module(msprofiler libmsprofiler.a ${ASCEND_ACL_DIR}) | find_module(msprofiler libmsprofiler.a ${ASCEND_ACL_DIR}) | ||||
#mdc没有 dc在driver/lib64/driver | |||||
find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}/driver) | find_module(ascend_hal_stub libascend_hal.so ${ASCEND_DRIVER_DIR}/driver) | ||||
find_module(ascendcl_static libascendcl.a ${ASCEND_ACL_DIR}) | |||||
#find_module(ascendcl_static libascendcl.a ${ASCEND_ACL_DIR}) | |||||
else() | else() | ||||
message(FATAL_ERROR "PLATFORM param is invalid, should be train or inference, build terminated") | message(FATAL_ERROR "PLATFORM param is invalid, should be train or inference, build terminated") | ||||
endif() | endif() | ||||
@@ -95,6 +95,9 @@ checkopts() | |||||
} | } | ||||
checkopts "$@" | checkopts "$@" | ||||
git submodule update --init metadef | |||||
git submodule update --init parser | |||||
mk_dir() { | mk_dir() { | ||||
local create_dir="$1" # the target to make | local create_dir="$1" # the target to make | ||||
@@ -134,7 +137,7 @@ build_graphengine() | |||||
echo "execute command: cmake ${CMAKE_ARGS} .. failed." | echo "execute command: cmake ${CMAKE_ARGS} .. failed." | ||||
return 1 | return 1 | ||||
fi | fi | ||||
COMMON_TARGET="ge_common engine fmk_parser parser_common _caffe_parser fmk_onnx_parser graph register " | |||||
COMMON_TARGET="ge_common engine fmk_parser parser_common _caffe_parser fmk_onnx_parser graph register engine_conf.json optimizer_priority.pbtxt " | |||||
TARGET=${COMMON_TARGET} | TARGET=${COMMON_TARGET} | ||||
if [ "x${PLATFORM}" = "xtrain" ] | if [ "x${PLATFORM}" = "xtrain" ] | ||||
then | then | ||||
@@ -211,14 +214,80 @@ generate_package() | |||||
cd "${BASEPATH}" | cd "${BASEPATH}" | ||||
GRAPHENGINE_LIB_PATH="lib" | GRAPHENGINE_LIB_PATH="lib" | ||||
ACL_PATH="acllib/lib64" | |||||
FWK_PATH="fwkacllib/lib64" | |||||
ATC_PATH="atc/lib64" | |||||
ATC_BIN_PATH="atc/bin" | |||||
NNENGINE_PATH="plugin/nnengine/ge_config" | |||||
OPSKERNEL_PATH="plugin/opskernel" | |||||
ATC_LIB=("libc_sec.so" "libge_common.so" "libge_compiler.so" "libgraph.so" "libregister.so") | |||||
FWK_LIB=("libge_common.so" "libge_runner.so" "libgraph.so" "libregister.so") | |||||
PLUGIN_OPSKERNEL=("libge_local_engine.so" "libge_local_opskernel_builder.so" "libhost_cpu_engine.so" "libhost_cpu_opskernel_builder.so" "optimizer_priority.pbtxt") | |||||
PARSER_LIB=("lib_caffe_parser.so" "libfmk_onnx_parser.so" "libfmk_parser.so" "libparser_common.so") | |||||
rm -rf ${OUTPUT_PATH:?}/${FWK_PATH}/ | |||||
rm -rf ${OUTPUT_PATH:?}/${ACL_PATH}/ | |||||
rm -rf ${OUTPUT_PATH:?}/${ATC_PATH}/ | |||||
rm -rf ${OUTPUT_PATH:?}/${ATC_BIN_PATH}/ | |||||
mk_dir "${OUTPUT_PATH}/${FWK_PATH}/${NNENGINE_PATH}" | |||||
mk_dir "${OUTPUT_PATH}/${FWK_PATH}/${OPSKERNEL_PATH}" | |||||
mk_dir "${OUTPUT_PATH}/${ATC_PATH}/${NNENGINE_PATH}" | |||||
mk_dir "${OUTPUT_PATH}/${ATC_PATH}/${OPSKERNEL_PATH}" | |||||
mk_dir "${OUTPUT_PATH}/${ACL_PATH}" | |||||
mk_dir "${OUTPUT_PATH}/${ATC_BIN_PATH}" | |||||
cd "${OUTPUT_PATH}" | cd "${OUTPUT_PATH}" | ||||
find ./ -name graphengine_lib.tar -exec rm {} \; | find ./ -name graphengine_lib.tar -exec rm {} \; | ||||
find ./bin -name atc -exec cp {} "${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH}" \; | |||||
cp ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH}/engine_conf.json ${OUTPUT_PATH}/${FWK_PATH}/${NNENGINE_PATH} | |||||
cp ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH}/engine_conf.json ${OUTPUT_PATH}/${ATC_PATH}/${NNENGINE_PATH} | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name libengine.so -exec cp -f {} ${OUTPUT_PATH}/${FWK_PATH}/${NNENGINE_PATH}/../ \; | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name libengine.so -exec cp -f {} ${OUTPUT_PATH}/${ATC_PATH}/${NNENGINE_PATH}/../ \; | |||||
MAX_DEPTH=1 | |||||
if [ "x${PLATFORM}" = "xall" ] || [ "x${PLATFORM}" = "xinference" ] | |||||
then | |||||
MAX_DEPTH=2 | |||||
fi | |||||
for lib in "${PLUGIN_OPSKERNEL[@]}"; | |||||
do | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth ${MAX_DEPTH} -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${FWK_PATH}/${OPSKERNEL_PATH} \; | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth ${MAX_DEPTH} -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${ATC_PATH}/${OPSKERNEL_PATH} \; | |||||
done | |||||
for lib in "${PARSER_LIB[@]}"; | |||||
do | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${FWK_PATH} \; | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${ATC_PATH} \; | |||||
done | |||||
for lib in "${FWK_LIB[@]}"; | |||||
do | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${FWK_PATH} \; | |||||
done | |||||
for lib in "${ATC_LIB[@]}"; | |||||
do | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name "$lib" -exec cp -f {} ${OUTPUT_PATH}/${ATC_PATH} \; | |||||
done | |||||
find ./bin -name atc -exec cp {} "${OUTPUT_PATH}/${ATC_BIN_PATH}" \; | |||||
find ${OUTPUT_PATH}/${GRAPHENGINE_LIB_PATH} -maxdepth 1 -name "libascendcl.so" -exec cp -f {} ${OUTPUT_PATH}/${ACL_PATH} \; | |||||
tar -cf graphengine_lib.tar "${GRAPHENGINE_LIB_PATH}" | |||||
if [ "x${PLATFORM}" = "xtrain" ] | |||||
then | |||||
tar -cf graphengine_lib.tar fwkacllib | |||||
elif [ "x${PLATFORM}" = "xinference" ] | |||||
then | |||||
tar -cf graphengine_lib.tar acllib atc | |||||
elif [ "x${PLATFORM}" = "xall" ] | |||||
then | |||||
tar -cf graphengine_lib.tar fwkacllib acllib atc | |||||
fi | |||||
} | } | ||||
if [[ "X$ENABLE_GE_UT" = "Xoff" ]]; then | if [[ "X$ENABLE_GE_UT" = "Xoff" ]]; then | ||||
@@ -88,6 +88,7 @@ set(TRAIN_SRC_LIST | |||||
"graph/load/new_model_manager/model_utils.cc" | "graph/load/new_model_manager/model_utils.cc" | ||||
"graph/load/new_model_manager/aipp_utils.cc" | "graph/load/new_model_manager/aipp_utils.cc" | ||||
"graph/load/new_model_manager/task_info/end_graph_task_info.cc" | "graph/load/new_model_manager/task_info/end_graph_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/model_exit_task_info.cc" | |||||
"graph/load/new_model_manager/task_info/event_record_task_info.cc" | "graph/load/new_model_manager/task_info/event_record_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/event_wait_task_info.cc" | "graph/load/new_model_manager/task_info/event_wait_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/fusion_start_task_info.cc" | "graph/load/new_model_manager/task_info/fusion_start_task_info.cc" | ||||
@@ -137,6 +138,7 @@ set(TRAIN_SRC_LIST | |||||
"graph/passes/atomic_addr_clean_pass.cc" | "graph/passes/atomic_addr_clean_pass.cc" | ||||
"graph/passes/mark_same_addr_pass.cc" | "graph/passes/mark_same_addr_pass.cc" | ||||
"graph/passes/mark_graph_unknown_status_pass.cc" | "graph/passes/mark_graph_unknown_status_pass.cc" | ||||
"graph/passes/mark_agnostic_pass.cc" | |||||
"graph/partition/dynamic_shape_partition.cc" | "graph/partition/dynamic_shape_partition.cc" | ||||
"graph/partition/stage_partition.cc" | "graph/partition/stage_partition.cc" | ||||
"graph/passes/base_pass.cc" | "graph/passes/base_pass.cc" | ||||
@@ -489,6 +491,7 @@ set(INFER_SRC_LIST | |||||
"graph/passes/atomic_addr_clean_pass.cc" | "graph/passes/atomic_addr_clean_pass.cc" | ||||
"graph/passes/mark_same_addr_pass.cc" | "graph/passes/mark_same_addr_pass.cc" | ||||
"graph/passes/mark_graph_unknown_status_pass.cc" | "graph/passes/mark_graph_unknown_status_pass.cc" | ||||
"graph/passes/mark_agnostic_pass.cc" | |||||
"graph/common/omg_util.cc" | "graph/common/omg_util.cc" | ||||
"graph/common/bcast.cc" | "graph/common/bcast.cc" | ||||
"graph/common/local_context.cc" | "graph/common/local_context.cc" | ||||
@@ -623,6 +626,7 @@ set(INFER_SRC_LIST | |||||
"graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | "graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/stream_switchn_task_info.cc" | "graph/load/new_model_manager/task_info/stream_switchn_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/end_graph_task_info.cc" | "graph/load/new_model_manager/task_info/end_graph_task_info.cc" | ||||
"graph/load/new_model_manager/task_info/model_exit_task_info.cc" | |||||
"graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | "graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | ||||
"graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | "graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | ||||
"single_op/task/op_task.cc" | "single_op/task/op_task.cc" | ||||
@@ -688,7 +692,7 @@ target_include_directories(ge_compiler PRIVATE | |||||
${GE_CODE_DIR}/../inc/external | ${GE_CODE_DIR}/../inc/external | ||||
${GE_CODE_DIR}/../inc/cce | ${GE_CODE_DIR}/../inc/cce | ||||
${GE_CODE_DIR}/../toolchain/ide/ide-daemon/external | ${GE_CODE_DIR}/../toolchain/ide/ide-daemon/external | ||||
#### blue zone | |||||
#### blue zone #### | |||||
${ASCEND_DIR}/driver/include | ${ASCEND_DIR}/driver/include | ||||
${ASCEND_DIR}/fwkacllib/include | ${ASCEND_DIR}/fwkacllib/include | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | ${GE_CODE_DIR}/third_party/fwkacllib/inc | ||||
@@ -717,10 +721,24 @@ target_link_libraries(ge_compiler | |||||
############ libascendcl.so ############ | ############ libascendcl.so ############ | ||||
file(GENERATE OUTPUT ${CMAKE_BINARY_DIR}/dummy.c CONTENT "") | file(GENERATE OUTPUT ${CMAKE_BINARY_DIR}/dummy.c CONTENT "") | ||||
add_library(dummy_obj OBJECT ${CMAKE_BINARY_DIR}/dummy.c) | |||||
set(DUMMY_OBJ $<TARGET_OBJECTS:dummy_obj>) | |||||
#add_library(dummy_obj OBJECT ${CMAKE_BINARY_DIR}/dummy.c) | |||||
#set(DUMMY_OBJ $<TARGET_OBJECTS:dummy_obj>) | |||||
file(MAKE_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}/ascendcl_object) | |||||
add_library(opensrc_ascendcl SHARED ${DUMMY_OBJ}) | |||||
if(EXISTS ${STATIC_ACL_LIB}/libascendcl.a) | |||||
execute_process( | |||||
COMMAND ar x ${STATIC_ACL_LIB}/libascendcl.a | |||||
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}/ascendcl_object | |||||
) | |||||
file(GLOB OBJECT_LIST ${CMAKE_CURRENT_BINARY_DIR}/ascendcl_object/*.o) | |||||
else() | |||||
set(OBJECT_LIST ${CMAKE_BINARY_DIR}/dummy.c) | |||||
endif() | |||||
add_library(opensrc_ascendcl SHARED | |||||
${OBJECT_LIST} | |||||
) | |||||
target_compile_options(opensrc_ascendcl PRIVATE | target_compile_options(opensrc_ascendcl PRIVATE | ||||
-O2 | -O2 | ||||
-fvisibility=hidden | -fvisibility=hidden | ||||
@@ -735,7 +753,6 @@ target_link_options(opensrc_ascendcl PRIVATE | |||||
target_link_libraries(opensrc_ascendcl PRIVATE | target_link_libraries(opensrc_ascendcl PRIVATE | ||||
-Wl,--whole-archive | -Wl,--whole-archive | ||||
ge_executor | ge_executor | ||||
ascendcl_static | |||||
ge_common_static | ge_common_static | ||||
graph_static | graph_static | ||||
protobuf_static | protobuf_static | ||||
@@ -879,7 +896,7 @@ add_custom_command( | |||||
set(INSTALL_BASE_DIR "") | set(INSTALL_BASE_DIR "") | ||||
set(INSTALL_LIBRARY_DIR lib) | set(INSTALL_LIBRARY_DIR lib) | ||||
install(TARGETS ge_runner ge_compiler OPTIONAL | |||||
install(TARGETS ge_runner ge_compiler opensrc_ascendcl OPTIONAL | |||||
LIBRARY DESTINATION ${INSTALL_LIBRARY_DIR} | LIBRARY DESTINATION ${INSTALL_LIBRARY_DIR} | ||||
) | ) | ||||
@@ -45,6 +45,9 @@ message AippOpParams { | |||||
// 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | // 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | ||||
uint32 related_input_rank = 2; | uint32 related_input_rank = 2; | ||||
// related_input_name is optional and the top name of data node which inserts aipp | |||||
string related_input_name = 6; | |||||
// input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | // input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | ||||
// 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | // 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | ||||
// 配置值 <= Data算子输出边的个数。 | // 配置值 <= Data算子输出边的个数。 | ||||
@@ -45,6 +45,9 @@ message AippOpParams { | |||||
// 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | // 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | ||||
uint32 related_input_rank = 2; | uint32 related_input_rank = 2; | ||||
// related_input_name is optional and the top name of data node which inserts aipp | |||||
string related_input_name = 6; | |||||
// input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | // input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | ||||
// 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | // 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | ||||
// 配置值 <= Data算子输出边的个数。 | // 配置值 <= Data算子输出边的个数。 | ||||
@@ -398,6 +398,7 @@ REGISTER_OPTYPE_DEFINE(MEMCPYASYNC, "MemcpyAsync"); | |||||
REGISTER_OPTYPE_DEFINE(MEMCPYADDRASYNC, "MemcpyAddrAsync"); | REGISTER_OPTYPE_DEFINE(MEMCPYADDRASYNC, "MemcpyAddrAsync"); | ||||
REGISTER_OPTYPE_DEFINE(STREAMMERGE, "StreamMerge"); | REGISTER_OPTYPE_DEFINE(STREAMMERGE, "StreamMerge"); | ||||
REGISTER_OPTYPE_DEFINE(ENDGRAPH, "EndGraph"); | REGISTER_OPTYPE_DEFINE(ENDGRAPH, "EndGraph"); | ||||
REGISTER_OPTYPE_DEFINE(MODELEXIT, "ModelExit"); | |||||
REGISTER_OPTYPE_DEFINE(SEND, "Send"); | REGISTER_OPTYPE_DEFINE(SEND, "Send"); | ||||
REGISTER_OPTYPE_DEFINE(RECV, "Recv"); | REGISTER_OPTYPE_DEFINE(RECV, "Recv"); | ||||
REGISTER_OPTYPE_DEFINE(ENDOFSEQUENCE, "EndOfSequence"); | REGISTER_OPTYPE_DEFINE(ENDOFSEQUENCE, "EndOfSequence"); | ||||
@@ -58,6 +58,7 @@ set(SRC_LIST | |||||
"../graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | "../graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | ||||
"../graph/load/new_model_manager/task_info/stream_switchn_task_info.cc" | "../graph/load/new_model_manager/task_info/stream_switchn_task_info.cc" | ||||
"../graph/load/new_model_manager/task_info/end_graph_task_info.cc" | "../graph/load/new_model_manager/task_info/end_graph_task_info.cc" | ||||
"../graph/load/new_model_manager/task_info/model_exit_task_info.cc" | |||||
"../graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | "../graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | ||||
"../graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | "../graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | ||||
"../opskernel_manager/ops_kernel_builder_manager.cc" | "../opskernel_manager/ops_kernel_builder_manager.cc" | ||||
@@ -91,7 +92,6 @@ target_include_directories(ge_executor PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -100,7 +100,9 @@ target_include_directories(ge_executor PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
${GE_CODE_DIR}/../inc/cce | |||||
${GE_CODE_DIR}/../inc/cce | |||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(ge_executor PRIVATE | target_link_libraries(ge_executor PRIVATE | ||||
@@ -1058,6 +1058,14 @@ ge::Status GeExecutor::ExecuteAsync(DynamicSingleOp *executor, const vector<GeTe | |||||
} | } | ||||
Status GeExecutor::ReleaseSingleOpResource(void *stream) { | Status GeExecutor::ReleaseSingleOpResource(void *stream) { | ||||
// get current context | |||||
rtContext_t rt_cur_ctx = nullptr; | |||||
auto rt_err = rtCtxGetCurrent(&rt_cur_ctx); | |||||
if (rt_err != RT_ERROR_NONE) { | |||||
GELOGE(RT_FAILED, "get current context failed, runtime result is %d", static_cast<int>(rt_err)); | |||||
return RT_FAILED; | |||||
} | |||||
ModelManager::GetInstance()->ClearAICPUSo(rt_cur_ctx); | |||||
return SingleOpManager::GetInstance().ReleaseResource(stream); | return SingleOpManager::GetInstance().ReleaseResource(stream); | ||||
} | } | ||||
@@ -48,6 +48,7 @@ local_ge_executor_src_files := \ | |||||
../graph/load/new_model_manager/task_info/stream_switch_task_info.cc \ | ../graph/load/new_model_manager/task_info/stream_switch_task_info.cc \ | ||||
../graph/load/new_model_manager/task_info/stream_switchn_task_info.cc \ | ../graph/load/new_model_manager/task_info/stream_switchn_task_info.cc \ | ||||
../graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | ../graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | ||||
../graph/load/new_model_manager/task_info/model_exit_task_info.cc \ | |||||
../graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc \ | ../graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc \ | ||||
../graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc \ | ../graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc \ | ||||
../opskernel_manager/ops_kernel_builder_manager.cc \ | ../opskernel_manager/ops_kernel_builder_manager.cc \ | ||||
@@ -45,6 +45,9 @@ message AippOpParams { | |||||
// 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | // 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | ||||
uint32 related_input_rank = 2; | uint32 related_input_rank = 2; | ||||
// related_input_name is optional and the top name of data node which inserts aipp | |||||
string related_input_name = 6; | |||||
// input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | // input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | ||||
// 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | // 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | ||||
// 配置值 <= Data算子输出边的个数。 | // 配置值 <= Data算子输出边的个数。 | ||||
@@ -109,6 +109,7 @@ OMG_HOST_SRC_FILES := \ | |||||
graph/passes/atomic_addr_clean_pass.cc \ | graph/passes/atomic_addr_clean_pass.cc \ | ||||
graph/passes/mark_same_addr_pass.cc \ | graph/passes/mark_same_addr_pass.cc \ | ||||
graph/passes/mark_graph_unknown_status_pass.cc \ | graph/passes/mark_graph_unknown_status_pass.cc \ | ||||
graph/passes/mark_agnostic_pass.cc \ | |||||
graph/common/omg_util.cc \ | graph/common/omg_util.cc \ | ||||
graph/common/bcast.cc \ | graph/common/bcast.cc \ | ||||
graph/common/local_context.cc \ | graph/common/local_context.cc \ | ||||
@@ -248,6 +249,7 @@ OME_HOST_SRC_FILES := \ | |||||
graph/load/new_model_manager/task_info/stream_switch_task_info.cc \ | graph/load/new_model_manager/task_info/stream_switch_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/stream_switchn_task_info.cc \ | graph/load/new_model_manager/task_info/stream_switchn_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/model_exit_task_info.cc \ | |||||
graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc \ | graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc \ | ||||
graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc \ | graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc \ | ||||
single_op/task/op_task.cc \ | single_op/task/op_task.cc \ | ||||
@@ -34,7 +34,6 @@ target_include_directories(ge_local_engine PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -43,6 +42,8 @@ target_include_directories(ge_local_engine PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(ge_local_engine PRIVATE | target_link_libraries(ge_local_engine PRIVATE | ||||
@@ -74,7 +75,6 @@ target_include_directories(atc_ge_local_engine PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -83,6 +83,8 @@ target_include_directories(atc_ge_local_engine PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(atc_ge_local_engine PRIVATE | target_link_libraries(atc_ge_local_engine PRIVATE | ||||
@@ -115,7 +117,6 @@ target_include_directories(ge_local_opskernel_builder PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -124,6 +125,8 @@ target_include_directories(ge_local_opskernel_builder PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(ge_local_opskernel_builder PRIVATE | target_link_libraries(ge_local_opskernel_builder PRIVATE | ||||
@@ -150,7 +153,6 @@ target_include_directories(atc_ge_local_opskernel_builder PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -159,6 +161,8 @@ target_include_directories(atc_ge_local_opskernel_builder PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(atc_ge_local_opskernel_builder PRIVATE | target_link_libraries(atc_ge_local_opskernel_builder PRIVATE | ||||
@@ -190,7 +194,6 @@ target_include_directories(ge_local_opskernel_builder_static PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -199,6 +202,8 @@ target_include_directories(ge_local_opskernel_builder_static PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(ge_local_opskernel_builder_static PRIVATE | target_link_libraries(ge_local_opskernel_builder_static PRIVATE | ||||
@@ -61,6 +61,7 @@ LIBGE_LOCAL_SRC_FILES := \ | |||||
graph/load/new_model_manager/model_utils.cc \ | graph/load/new_model_manager/model_utils.cc \ | ||||
graph/load/new_model_manager/aipp_utils.cc \ | graph/load/new_model_manager/aipp_utils.cc \ | ||||
graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | graph/load/new_model_manager/task_info/end_graph_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/model_exit_task_info.cc \ | |||||
graph/load/new_model_manager/task_info/event_record_task_info.cc \ | graph/load/new_model_manager/task_info/event_record_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/event_wait_task_info.cc \ | graph/load/new_model_manager/task_info/event_wait_task_info.cc \ | ||||
graph/load/new_model_manager/task_info/fusion_start_task_info.cc \ | graph/load/new_model_manager/task_info/fusion_start_task_info.cc \ | ||||
@@ -110,6 +111,7 @@ LIBGE_LOCAL_SRC_FILES := \ | |||||
graph/passes/atomic_addr_clean_pass.cc \ | graph/passes/atomic_addr_clean_pass.cc \ | ||||
graph/passes/mark_same_addr_pass.cc \ | graph/passes/mark_same_addr_pass.cc \ | ||||
graph/passes/mark_graph_unknown_status_pass.cc \ | graph/passes/mark_graph_unknown_status_pass.cc \ | ||||
graph/passes/mark_agnostic_pass.cc \ | |||||
graph/partition/dynamic_shape_partition.cc \ | graph/partition/dynamic_shape_partition.cc \ | ||||
graph/partition/stage_partition.cc \ | graph/partition/stage_partition.cc \ | ||||
graph/passes/base_pass.cc \ | graph/passes/base_pass.cc \ | ||||
@@ -31,7 +31,8 @@ target_include_directories(ge_memory PRIVATE | |||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) |
@@ -866,12 +866,10 @@ MemoryBlock *BlockMemAssigner::ApplyMemory(size_t block_size, size_t real_size, | |||||
string ge_disable_reuse_mem_env = "0"; | string ge_disable_reuse_mem_env = "0"; | ||||
(void)ge::GetContext().GetOption(OPTION_EXEC_DISABLE_REUSED_MEMORY, ge_disable_reuse_mem_env); | (void)ge::GetContext().GetOption(OPTION_EXEC_DISABLE_REUSED_MEMORY, ge_disable_reuse_mem_env); | ||||
if (ge_disable_reuse_mem_env != "1") { | if (ge_disable_reuse_mem_env != "1") { | ||||
bool reuse_mem_flag = !((workspace_reuse_flag.size() > out_index) && !workspace_reuse_flag[out_index]); | |||||
bool reuse_mem_flag = (mem_type == kOutput) ? IsPreReuse(n, out_index) : | |||||
!((workspace_reuse_flag.size() > out_index) && !workspace_reuse_flag[out_index]); | |||||
is_reuse_memory = !node_op_desc->HasAttr(kL2FusionDynamicConvergeOp) && | is_reuse_memory = !node_op_desc->HasAttr(kL2FusionDynamicConvergeOp) && | ||||
!node_op_desc->HasAttr(kOpNoReuseMem) && | |||||
reuse_mem_flag && | |||||
is_op_reuse_mem && | |||||
(IsPreReuse(n, out_index)); | |||||
!node_op_desc->HasAttr(kOpNoReuseMem) && reuse_mem_flag && is_op_reuse_mem; | |||||
auto stream_id = node_op_desc->GetStreamId(); | auto stream_id = node_op_desc->GetStreamId(); | ||||
if (is_reuse_memory && !continuous && !reusable_blocks_[memory_type].empty()) { | if (is_reuse_memory && !continuous && !reusable_blocks_[memory_type].empty()) { | ||||
for (auto it = reusable_blocks_[memory_type][stream_id].begin(); | for (auto it = reusable_blocks_[memory_type][stream_id].begin(); | ||||
@@ -236,6 +236,7 @@ ModelManager::~ModelManager() { | |||||
std::lock_guard<std::mutex> lock(map_mutex_); | std::lock_guard<std::mutex> lock(map_mutex_); | ||||
model_map_.clear(); | model_map_.clear(); | ||||
model_aicpu_kernel_.clear(); | model_aicpu_kernel_.clear(); | ||||
cust_aicpu_so_.clear(); | |||||
GE_IF_BOOL_EXEC(device_count > 0, GE_CHK_RT(rtDeviceReset(0))); | GE_IF_BOOL_EXEC(device_count > 0, GE_CHK_RT(rtDeviceReset(0))); | ||||
} | } | ||||
@@ -399,7 +400,6 @@ Status ModelManager::Unload(uint32_t model_id) { | |||||
} | } | ||||
std::lock_guard<std::mutex> lock(exeception_infos_mutex_); | std::lock_guard<std::mutex> lock(exeception_infos_mutex_); | ||||
exception_infos_.clear(); | exception_infos_.clear(); | ||||
cust_aicpu_so_.clear(); | |||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
@@ -1096,19 +1096,51 @@ Status ModelManager::CreateAicpuSession(uint64_t session_id) { | |||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
Status ModelManager::LoadCustAicpuSo(const OpDescPtr op_desc, string so_name) { | |||||
Status ModelManager::LoadCustAicpuSo(const OpDescPtr op_desc, const string &so_name) { | |||||
GELOGI("LoadCustAicpuSo in, op name %s, so_name %s.", op_desc->GetName().c_str(), so_name.c_str()); | |||||
std::lock_guard<std::mutex> lock(cust_aicpu_mutex_); | std::lock_guard<std::mutex> lock(cust_aicpu_mutex_); | ||||
auto it = cust_aicpu_so_.find(so_name); | |||||
// get current context | |||||
rtContext_t rt_cur_ctx = nullptr; | |||||
auto rt_err = rtCtxGetCurrent(&rt_cur_ctx); | |||||
if (rt_err != RT_ERROR_NONE) { | |||||
GELOGE(RT_FAILED, "get current context failed, runtime result is %d", static_cast<int>(rt_err)); | |||||
return RT_FAILED; | |||||
} | |||||
// use current context as resource key instead | |||||
uintptr_t resource_id = reinterpret_cast<uintptr_t>(rt_cur_ctx); | |||||
auto it = cust_aicpu_so_.find(resource_id); | |||||
if (it == cust_aicpu_so_.end()) { | if (it == cust_aicpu_so_.end()) { | ||||
GE_CHK_STATUS_RET(LaunchCustAicpuSo(op_desc, so_name), "LaunchCustAicpuSo failed. op name %s, so_name %s", | GE_CHK_STATUS_RET(LaunchCustAicpuSo(op_desc, so_name), "LaunchCustAicpuSo failed. op name %s, so_name %s", | ||||
op_desc->GetName().c_str(), so_name.c_str()); | op_desc->GetName().c_str(), so_name.c_str()); | ||||
(void)cust_aicpu_so_.insert(so_name); | |||||
GELOGI("LaunchCustAicpuSo op name %s, so_name %s.", op_desc->GetName().c_str(), so_name.c_str()); | |||||
std::set<string> so_name_set; | |||||
so_name_set.insert(so_name); | |||||
cust_aicpu_so_[resource_id] = so_name_set; | |||||
GELOGI("LoadCustAicpuSo new aicpu so resource_id %lu.", resource_id); | |||||
return SUCCESS; | |||||
} | |||||
auto it_so_name = it->second.find(so_name); | |||||
if (it_so_name == it->second.end()) { | |||||
GE_CHK_STATUS_RET(LaunchCustAicpuSo(op_desc, so_name), "LaunchCustAicpuSo failed. op name %s, so_name %s", | |||||
op_desc->GetName().c_str(), so_name.c_str()); | |||||
it->second.insert(so_name); | |||||
GELOGI("LoadCustAicpuSo add aicpu so resource_id %lu.", resource_id); | |||||
} | |||||
return SUCCESS; | |||||
} | |||||
Status ModelManager::ClearAICPUSo(void *ctx) { | |||||
auto ctx_id = reinterpret_cast<uintptr_t>(ctx); | |||||
GELOGI("ClearAICPUSo in. resource id = 0x%lx", static_cast<uint64_t>(ctx_id)); | |||||
std::lock_guard<std::mutex> lock(cust_aicpu_mutex_); | |||||
auto it = cust_aicpu_so_.find(ctx_id); | |||||
if (it == cust_aicpu_so_.end()) { | |||||
return SUCCESS; | |||||
} | } | ||||
(void)cust_aicpu_so_.erase(it); | |||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
Status ModelManager::LaunchCustAicpuSo(const OpDescPtr op_desc, string so_name) { | |||||
Status ModelManager::LaunchCustAicpuSo(const OpDescPtr op_desc, const string &so_name) { | |||||
CustAICPUKernelPtr aicpu_kernel = op_desc->TryGetExtAttr(OP_EXTATTR_CUSTAICPU_KERNEL, CustAICPUKernelPtr()); | CustAICPUKernelPtr aicpu_kernel = op_desc->TryGetExtAttr(OP_EXTATTR_CUSTAICPU_KERNEL, CustAICPUKernelPtr()); | ||||
if (aicpu_kernel == nullptr) { | if (aicpu_kernel == nullptr) { | ||||
GELOGE(INTERNAL_ERROR, "cust aicpu op %s can't find kernel!", op_desc->GetName().c_str()); | GELOGE(INTERNAL_ERROR, "cust aicpu op %s can't find kernel!", op_desc->GetName().c_str()); | ||||
@@ -1140,6 +1172,7 @@ Status ModelManager::LaunchCustAicpuSo(const OpDescPtr op_desc, string so_name) | |||||
GE_CHK_RT(rtStreamCreate(&stream, 0)); | GE_CHK_RT(rtStreamCreate(&stream, 0)); | ||||
GE_CHK_RT(rtCpuKernelLaunch(nullptr, kLoadOpFromBuf, 1, args, args_size, nullptr, stream)); | GE_CHK_RT(rtCpuKernelLaunch(nullptr, kLoadOpFromBuf, 1, args, args_size, nullptr, stream)); | ||||
GELOGI("LaunchCustAicpuSo so buf len %u, so name len %u.", aicpu_data_length, so_name.size()); | |||||
status = rtStreamSynchronize(stream); | status = rtStreamSynchronize(stream); | ||||
if (status != RT_ERROR_NONE) { | if (status != RT_ERROR_NONE) { | ||||
GELOGE(RT_FAILED, "Call rt stream sync failed, status: 0x%x", status); | GELOGE(RT_FAILED, "Call rt stream sync failed, status: 0x%x", status); | ||||
@@ -270,9 +270,9 @@ class FMK_FUNC_HOST_VISIBILITY FMK_FUNC_DEV_VISIBILITY ModelManager { | |||||
ge::Status DestroyAicpuSessionForInfer(uint32_t model_id); | ge::Status DestroyAicpuSessionForInfer(uint32_t model_id); | ||||
ge::Status LoadCustAicpuSo(const OpDescPtr op_desc, string so_name); | |||||
ge::Status LaunchCustAicpuSo(const OpDescPtr op_desc, string so_name); | |||||
ge::Status LoadCustAicpuSo(const OpDescPtr op_desc, const string &so_name); | |||||
ge::Status LaunchCustAicpuSo(const OpDescPtr op_desc, const string &so_name); | |||||
ge::Status ClearAICPUSo(void *ctx); | |||||
ge::Status GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info); | ge::Status GetOrigInputInfo(uint32_t model_id, uint32_t index, OriginInputInfo &orig_input_info); | ||||
@@ -340,7 +340,7 @@ class FMK_FUNC_HOST_VISIBILITY FMK_FUNC_DEV_VISIBILITY ModelManager { | |||||
std::set<uint64_t> sess_ids_; | std::set<uint64_t> sess_ids_; | ||||
std::vector<rtExceptionInfo> exception_infos_; | std::vector<rtExceptionInfo> exception_infos_; | ||||
std::mutex cust_aicpu_mutex_; | std::mutex cust_aicpu_mutex_; | ||||
std::set<std::string> cust_aicpu_so_; | |||||
std::map<uintptr_t, std::set<std::string>> cust_aicpu_so_; | |||||
static DumpProperties dump_properties_; | static DumpProperties dump_properties_; | ||||
}; | }; | ||||
@@ -479,13 +479,15 @@ vector<void *> ModelUtils::GetWorkspaceDataAddrs(const RuntimeParam &model_param | |||||
ge::AttrUtils::GetListInt(op_desc, ATTR_NAME_WORKSPACE_TYPE_LIST, workspace_memory_type); | ge::AttrUtils::GetListInt(op_desc, ATTR_NAME_WORKSPACE_TYPE_LIST, workspace_memory_type); | ||||
for (size_t i = 0; i < v_workspace_bytes.size(); ++i) { | for (size_t i = 0; i < v_workspace_bytes.size(); ++i) { | ||||
// Temporary solution, the aicpu workspace of multiple images cannot be shared. | // Temporary solution, the aicpu workspace of multiple images cannot be shared. | ||||
if (has_workspace_reuse && i < workspace_reuse_flag.size() && !workspace_reuse_flag[i]) { | |||||
if (has_workspace_reuse && i < workspace_reuse_flag.size() | |||||
&& !workspace_reuse_flag[i] && !model_param.is_single_op) { | |||||
void *mem_addr = model_param.aicpu_mem_mall->Acquire(v_workspace_offset[i], v_workspace_bytes[i]); | void *mem_addr = model_param.aicpu_mem_mall->Acquire(v_workspace_offset[i], v_workspace_bytes[i]); | ||||
v_workspace_data_addr.push_back(mem_addr); | v_workspace_data_addr.push_back(mem_addr); | ||||
GELOGI( | GELOGI( | ||||
"[IMAS]GetWorkspaceDataAddrs graph_%u type[F] name[%s] aicpu workspace[%zu] offset[%ld] bytes[%ld] " | "[IMAS]GetWorkspaceDataAddrs graph_%u type[F] name[%s] aicpu workspace[%zu] offset[%ld] bytes[%ld] " | ||||
"memaddr[%p]", | "memaddr[%p]", | ||||
model_param.graph_id, op_desc->GetName().c_str(), i, v_workspace_offset[i], v_workspace_bytes[i], mem_addr); | model_param.graph_id, op_desc->GetName().c_str(), i, v_workspace_offset[i], v_workspace_bytes[i], mem_addr); | ||||
continue; | |||||
} else if (has_mem_type_workspace && workspace_memory_type[i] == RT_MEMORY_P2P_DDR) { | } else if (has_mem_type_workspace && workspace_memory_type[i] == RT_MEMORY_P2P_DDR) { | ||||
int64_t p2p_workspace_offset = v_workspace_offset[i]; | int64_t p2p_workspace_offset = v_workspace_offset[i]; | ||||
int64_t p2p_workspace_bytes = v_workspace_bytes[i]; | int64_t p2p_workspace_bytes = v_workspace_bytes[i]; | ||||
@@ -0,0 +1,54 @@ | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* | |||||
* Licensed under the Apache License, Version 2.0 (the "License"); | |||||
* you may not use this file except in compliance with the License. | |||||
* You may obtain a copy of the License at | |||||
* | |||||
* http://www.apache.org/licenses/LICENSE-2.0 | |||||
* | |||||
* Unless required by applicable law or agreed to in writing, software | |||||
* distributed under the License is distributed on an "AS IS" BASIS, | |||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||||
* See the License for the specific language governing permissions and | |||||
* limitations under the License. | |||||
*/ | |||||
#include "graph/load/new_model_manager/task_info/model_exit_task_info.h" | |||||
#include "common/properties_manager.h" | |||||
#include "framework/common/debug/ge_log.h" | |||||
#include "graph/load/new_model_manager/davinci_model.h" | |||||
namespace ge { | |||||
Status ModelExitTaskInfo::Init(const domi::TaskDef &task_def, DavinciModel *davinci_model) { | |||||
GELOGI("InitModelExitTaskInfo Init Start."); | |||||
if (davinci_model == nullptr) { | |||||
GELOGE(PARAM_INVALID, "davinci_model is null!"); | |||||
return PARAM_INVALID; | |||||
} | |||||
Status ret = SetStream(task_def.stream_id(), davinci_model->GetStreamList()); | |||||
if (ret != SUCCESS) { | |||||
GELOGE(ret, "SetStream fail, stream_id:%u", task_def.stream_id()); | |||||
return ret; | |||||
} | |||||
model_ = davinci_model->GetRtModelHandle(); | |||||
GELOGI("InitModelExitTaskInfo Init Success, model:%p, stream:%p", model_, stream_); | |||||
return SUCCESS; | |||||
} | |||||
Status ModelExitTaskInfo::Distribute() { | |||||
GELOGI("ModelExitTaskInfo Distribute Start."); | |||||
rtError_t rt_ret = rtModelExit(model_, stream_); | |||||
if (rt_ret != RT_ERROR_NONE) { | |||||
GELOGE(RT_FAILED, "Call rtModelExit failed, ret: 0x%x", rt_ret); | |||||
return RT_ERROR_TO_GE_STATUS(rt_ret); | |||||
} | |||||
GELOGI("ModelExitTaskInfo Distribute Success."); | |||||
return SUCCESS; | |||||
} | |||||
REGISTER_TASK_INFO(RT_MODEL_TASK_MODEL_EXIT, ModelExitTaskInfo); | |||||
} // namespace ge |
@@ -0,0 +1,37 @@ | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* | |||||
* Licensed under the Apache License, Version 2.0 (the "License"); | |||||
* you may not use this file except in compliance with the License. | |||||
* You may obtain a copy of the License at | |||||
* | |||||
* http://www.apache.org/licenses/LICENSE-2.0 | |||||
* | |||||
* Unless required by applicable law or agreed to in writing, software | |||||
* distributed under the License is distributed on an "AS IS" BASIS, | |||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||||
* See the License for the specific language governing permissions and | |||||
* limitations under the License. | |||||
*/ | |||||
#ifndef GE_GRAPH_LOAD_NEW_MODEL_MANAGER_TASK_INFO_MODEL_EXIT_TASK_INFO_H_ | |||||
#define GE_GRAPH_LOAD_NEW_MODEL_MANAGER_TASK_INFO_MODEL_EXIT_TASK_INFO_H_ | |||||
#include "graph/load/new_model_manager/task_info/task_info.h" | |||||
namespace ge { | |||||
class ModelExitTaskInfo : public TaskInfo { | |||||
public: | |||||
ModelExitTaskInfo() {} | |||||
~ModelExitTaskInfo() override { model_ = nullptr; } | |||||
Status Init(const domi::TaskDef &task_def, DavinciModel *davinci_model) override; | |||||
Status Distribute() override; | |||||
private: | |||||
rtModel_t model_{nullptr}; | |||||
}; | |||||
} // namespace ge | |||||
#endif // GE_GRAPH_LOAD_NEW_MODEL_MANAGER_TASK_INFO_MODEL_EXIT_TASK_INFO_H_ |
@@ -56,6 +56,7 @@ struct RuntimeParam { | |||||
uint32_t label_num = 0; | uint32_t label_num = 0; | ||||
uint64_t session_id = 0; | uint64_t session_id = 0; | ||||
uint32_t graph_id = 0; | uint32_t graph_id = 0; | ||||
bool is_single_op = false; | |||||
std::unique_ptr<TsMemMall> ts_mem_mall; | std::unique_ptr<TsMemMall> ts_mem_mall; | ||||
std::unique_ptr<TsMemMall> aicpu_mem_mall; | std::unique_ptr<TsMemMall> aicpu_mem_mall; | ||||
@@ -694,6 +694,7 @@ Status GraphManager::PreRun(const GraphNodePtr &graph_node, const std::vector<Ge | |||||
} | } | ||||
} | } | ||||
// BUILD_MODE_TUNING with BUILD_STEP_AFTER_MERGE no need PreRunOptimizeSubGraph. | // BUILD_MODE_TUNING with BUILD_STEP_AFTER_MERGE no need PreRunOptimizeSubGraph. | ||||
bool run_optimize_subgraph = | bool run_optimize_subgraph = | ||||
!((options_.build_mode == BUILD_MODE_TUNING) && (options_.build_step == BUILD_STEP_AFTER_MERGE)); | !((options_.build_mode == BUILD_MODE_TUNING) && (options_.build_step == BUILD_STEP_AFTER_MERGE)); | ||||
@@ -84,6 +84,22 @@ Status FlowCtrlPass::Run(ComputeGraphPtr compute_graph) { | |||||
return graph_change ? SUCCESS : NOT_CHANGED; | return graph_change ? SUCCESS : NOT_CHANGED; | ||||
} | } | ||||
bool FlowCtrlPass::CheckMultiDataSet(ComputeGraphPtr &compute_graph) { | |||||
int data_set_num = 0; | |||||
for (auto &node : compute_graph->GetDirectNode()) { | |||||
if (node == nullptr) { | |||||
continue; | |||||
} | |||||
string type; | |||||
bool is_found = AttrUtils::GetStr(node->GetOpDesc(), ATTR_NAME_FRAMEWORK_ORIGINAL_TYPE, type); | |||||
if (is_found && type == "IteratorV2") { | |||||
data_set_num++; | |||||
} | |||||
} | |||||
GELOGI("The ComputeGraph contain %d dataSet.", data_set_num); | |||||
return (data_set_num > 1) ? true : false; | |||||
} | |||||
NodePtr FlowCtrlPass::InsertOp(ComputeGraphPtr &compute_graph, const string &node_type, const string &node_name, | NodePtr FlowCtrlPass::InsertOp(ComputeGraphPtr &compute_graph, const string &node_type, const string &node_name, | ||||
const std::vector<GeTensorDesc> &input_list, | const std::vector<GeTensorDesc> &input_list, | ||||
const std::vector<GeTensorDesc> &output_list) { | const std::vector<GeTensorDesc> &output_list) { | ||||
@@ -310,12 +326,12 @@ Status FlowCtrlPass::CreateIterCtrlFalseBranch(ComputeGraphPtr &compute_graph, c | |||||
* loopCond | * loopCond | ||||
* | | * | | ||||
* v | * v | ||||
* switch --> Assign | |||||
* switch --> Assign --> active --> ModelExit | |||||
* ^ | * ^ | ||||
* | | * | | ||||
* loopReset | * loopReset | ||||
*/ | */ | ||||
// Insert Assign node | |||||
// Insert Assign node and ctrl edge | |||||
NodePtr assign_node = | NodePtr assign_node = | ||||
InsertAssignOp(compute_graph, ASSIGN, NODE_NAME_FLOWCTRL_LOOP_ASSIGN, loop_cond_node, loop_reset_node); | InsertAssignOp(compute_graph, ASSIGN, NODE_NAME_FLOWCTRL_LOOP_ASSIGN, loop_cond_node, loop_reset_node); | ||||
if (assign_node == nullptr || switch_node == nullptr) { | if (assign_node == nullptr || switch_node == nullptr) { | ||||
@@ -325,13 +341,49 @@ Status FlowCtrlPass::CreateIterCtrlFalseBranch(ComputeGraphPtr &compute_graph, c | |||||
GE_CHK_STATUS_RET(SetStreamLabel(assign_node, switch_node->GetName()), "set stream label failed"); | GE_CHK_STATUS_RET(SetStreamLabel(assign_node, switch_node->GetName()), "set stream label failed"); | ||||
// 3. Insert ctrl edges | |||||
graphStatus add_ret = GraphUtils::AddEdge(switch_node->GetOutControlAnchor(), assign_node->GetInControlAnchor()); | graphStatus add_ret = GraphUtils::AddEdge(switch_node->GetOutControlAnchor(), assign_node->GetInControlAnchor()); | ||||
if (add_ret != GRAPH_SUCCESS) { | if (add_ret != GRAPH_SUCCESS) { | ||||
GELOGE(FAILED, "Add switch_node to assign_node ctrl edge failed, add_ret=%u.", add_ret); | GELOGE(FAILED, "Add switch_node to assign_node ctrl edge failed, add_ret=%u.", add_ret); | ||||
return FAILED; | return FAILED; | ||||
} | } | ||||
if (CheckMultiDataSet(compute_graph)) { | |||||
GELOGI("Multi dataSae exist, model_exit node is need."); | |||||
// 2. Insert active node and add ctrl edge | |||||
string active_name = switch_node->GetName() + "_StreamExitActive"; | |||||
NodePtr active_node = InsertOp(compute_graph, STREAMACTIVE, active_name, {}, {}); | |||||
if (active_node == nullptr) { | |||||
GELOGE(FAILED, "Insert stream active node:%s for IterCtrlTrueStream failed.", active_name.c_str()); | |||||
return FAILED; | |||||
} | |||||
GE_CHK_STATUS_RET(SetStreamLabel(active_node, switch_node->GetName()), "set stream label failed"); | |||||
GE_IF_BOOL_EXEC(!AttrUtils::SetBool(active_node->GetOpDesc(), ATTR_NAME_IS_LOOP_ACTIVE, true), | |||||
DOMI_LOGE("set ATTR_NAME_IS_LOOP_ACTIVE failed"); return FAILED); | |||||
string model_exit_name = switch_node->GetName() + "_ModelExit"; | |||||
GE_CHK_STATUS_RET(SetActiveLabelList(active_node, { model_exit_name }), "set active label list failed"); | |||||
add_ret = GraphUtils::AddEdge(assign_node->GetOutControlAnchor(), active_node->GetInControlAnchor()); | |||||
if (add_ret != GRAPH_SUCCESS) { | |||||
GELOGE(FAILED, "Add assign_node to active_node ctrl edge failed, add_ret=%u.", add_ret); | |||||
return FAILED; | |||||
} | |||||
// 3. Insert model exit node and add ctrl edge | |||||
NodePtr model_exit_node = InsertOp(compute_graph, MODELEXIT, model_exit_name, {}, {}); | |||||
if (model_exit_node == nullptr) { | |||||
GELOGE(FAILED, "Insert model_exit node:%s for IterCtrlTrueStream failed.", model_exit_name.c_str()); | |||||
return FAILED; | |||||
} | |||||
GE_CHK_STATUS_RET(SetStreamLabel(model_exit_node, model_exit_name), "set stream label failed"); | |||||
add_ret = GraphUtils::AddEdge(active_node->GetOutControlAnchor(), model_exit_node->GetInControlAnchor()); | |||||
if (add_ret != GRAPH_SUCCESS) { | |||||
GELOGE(FAILED, "Add active_node to model_exit_node ctrl edge failed, add_ret=%u.", add_ret); | |||||
return FAILED; | |||||
} | |||||
} | |||||
GELOGI("CreateIterCtrlFalseBranch success."); | GELOGI("CreateIterCtrlFalseBranch success."); | ||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
@@ -134,6 +134,14 @@ class FlowCtrlPass : public GraphPass { | |||||
/// Other: failed | /// Other: failed | ||||
/// | /// | ||||
Status AddSpecialNodeIteratorCtrl(ComputeGraphPtr &compute_graph, NodePtr &loop_after_node); | Status AddSpecialNodeIteratorCtrl(ComputeGraphPtr &compute_graph, NodePtr &loop_after_node); | ||||
/// | |||||
/// add special iterator ctrl nodes(small cycle). | |||||
/// @param compute_graph graph | |||||
/// @return true: two or more dataSet exist | |||||
/// false: only one dataSet exist | |||||
/// | |||||
bool CheckMultiDataSet(ComputeGraphPtr &compute_graph); | |||||
}; | }; | ||||
} // namespace ge | } // namespace ge | ||||
@@ -15,20 +15,40 @@ | |||||
*/ | */ | ||||
#include "graph/passes/mark_agnostic_pass.h" | #include "graph/passes/mark_agnostic_pass.h" | ||||
#include "utils/node_utils.h" | |||||
#include "graph/utils/node_utils.h" | |||||
namespace ge { | namespace ge { | ||||
Status MarkAgnosticPass::Run(ComputeGraphPtr graph) { | Status MarkAgnosticPass::Run(ComputeGraphPtr graph) { | ||||
for (const auto &node : graph->GetDirectNode()) { | for (const auto &node : graph->GetDirectNode()) { | ||||
auto node_type = NodeUtils::GetNodeType(*node); | auto node_type = NodeUtils::GetNodeType(*node); | ||||
if (node_type == SWITCH || node_type == REFSWITCH || node_type == SWITCHN) { | if (node_type == SWITCH || node_type == REFSWITCH || node_type == SWITCHN) { | ||||
GELOGD("Mark format agnostic for switch ndoe %s", node->GetName().c_str()); | |||||
GELOGD("Mark format agnostic and continuous for switch node %s", node->GetName().c_str()); | |||||
const OpDescPtr op_desc = node->GetOpDesc(); | |||||
const GeTensorDescPtr op_tensor = op_desc->MutableInputDesc(0); | |||||
if (op_tensor == nullptr) { | |||||
GELOGD("Op: %s, Index:0,has no input", node->GetName().c_str()); | |||||
continue; | |||||
} | |||||
AttrUtils::SetInt(op_tensor, "_format_continuous", 1); | |||||
AttrUtils::SetInt(node->GetOpDesc(), "_format_agnostic", 1); | |||||
AttrUtils::SetListInt(node->GetOpDesc(), "_format_agnostic_except_input", std::vector<int64_t>({1})); | |||||
continue; | |||||
} | |||||
if (node_type == IDENTITY) { | |||||
GELOGD("Mark format agnostic for identity node %s", node->GetName().c_str()); | |||||
AttrUtils::SetInt(node->GetOpDesc(), "_format_agnostic", 1); | AttrUtils::SetInt(node->GetOpDesc(), "_format_agnostic", 1); | ||||
AttrUtils::SetListInt(node->GetOpDesc(), "_format_agnostic_except_input", std::vector<int64_t>({1})); | AttrUtils::SetListInt(node->GetOpDesc(), "_format_agnostic_except_input", std::vector<int64_t>({1})); | ||||
continue; | continue; | ||||
} | } | ||||
if (node_type == MERGE || node_type == REFMERGE) { | if (node_type == MERGE || node_type == REFMERGE) { | ||||
GELOGD("Mark format agnostic for merge node %s", node->GetName().c_str()); | |||||
GELOGD("Mark format agnostic and continuous for merge node %s", node->GetName().c_str()); | |||||
const OpDescPtr op_desc = node->GetOpDesc(); | |||||
const GeTensorDescPtr op_tensor = op_desc->MutableOutputDesc(0); | |||||
if (op_tensor == nullptr) { | |||||
GELOGD("Op: %s, Index:0,has no output", node->GetName().c_str()); | |||||
continue; | |||||
} | |||||
AttrUtils::SetInt(op_tensor, "_format_continuous", 1); | |||||
AttrUtils::SetInt(node->GetOpDesc(), "_format_agnostic", 1); | AttrUtils::SetInt(node->GetOpDesc(), "_format_agnostic", 1); | ||||
AttrUtils::SetListInt(node->GetOpDesc(), "_format_agnostic_except_output", std::vector<int64_t>({1})); | AttrUtils::SetListInt(node->GetOpDesc(), "_format_agnostic_except_output", std::vector<int64_t>({1})); | ||||
continue; | continue; | ||||
@@ -36,4 +56,4 @@ Status MarkAgnosticPass::Run(ComputeGraphPtr graph) { | |||||
} | } | ||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
} | |||||
} // namespace ge |
@@ -130,6 +130,14 @@ graphStatus TransOpWithoutReshapeFusionPass::GetSubGraphNodesInfo() { | |||||
sub_graph_has_reshape_node[i] = true; | sub_graph_has_reshape_node[i] = true; | ||||
break; | break; | ||||
} | } | ||||
if (in_node->GetType() == TRANSPOSE || in_node->GetType() == TRANSPOSED) { | |||||
auto input_format = in_node->GetOpDesc()->GetInputDescPtr(0)->GetFormat(); | |||||
auto output_format = in_node->GetOpDesc()->GetOutputDescPtr(0)->GetFormat(); | |||||
if (input_format == output_format) { | |||||
sub_graph_has_reshape_node[i] = true; | |||||
break; | |||||
} | |||||
} | |||||
auto out_anchor = iter->first; | auto out_anchor = iter->first; | ||||
GE_CHECK_NOTNULL(out_anchor); | GE_CHECK_NOTNULL(out_anchor); | ||||
@@ -46,6 +46,15 @@ Status TransposeTransDataPass::Run(NodePtr &node) { | |||||
if (op_desc->GetType() != TRANSPOSED) { | if (op_desc->GetType() != TRANSPOSED) { | ||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
auto input_format = op_desc->GetInputDescPtr(0)->GetFormat(); | |||||
auto output_format = op_desc->GetOutputDescPtr(0)->GetFormat(); | |||||
if (input_format == output_format) { | |||||
GELOGW("Node %s input format is %s, output format is %s, should not happend. Ignore pass.", | |||||
op_desc->GetName().c_str(), | |||||
TypeUtils::FormatToSerialString(input_format).c_str(), | |||||
TypeUtils::FormatToSerialString(output_format).c_str()); | |||||
return SUCCESS; | |||||
} | |||||
if (CheckOneInAndOneOutDataAnchor(node) != SUCCESS) { | if (CheckOneInAndOneOutDataAnchor(node) != SUCCESS) { | ||||
return FAILED; | return FAILED; | ||||
} | } | ||||
@@ -92,6 +92,7 @@ | |||||
#include "graph/passes/unused_op_remove_pass.h" | #include "graph/passes/unused_op_remove_pass.h" | ||||
#include "graph/passes/var_is_initialized_op_pass.h" | #include "graph/passes/var_is_initialized_op_pass.h" | ||||
#include "graph/passes/variable_prepare_op_pass.h" | #include "graph/passes/variable_prepare_op_pass.h" | ||||
#include "graph/passes/mark_agnostic_pass.h" | |||||
#include "graph/preprocess/insert_op/util_insert_aipp_op.h" | #include "graph/preprocess/insert_op/util_insert_aipp_op.h" | ||||
#include "graph/types.h" | #include "graph/types.h" | ||||
#include "graph/utils/tensor_utils.h" | #include "graph/utils/tensor_utils.h" | ||||
@@ -1626,6 +1627,7 @@ Status GraphPrepare::PrepareOptimize() { | |||||
try { | try { | ||||
(void)original_graph_passes.AddPass("PrepareOptimize::ShapeOperateOpRemovePass", new ShapeOperateOpRemovePass); | (void)original_graph_passes.AddPass("PrepareOptimize::ShapeOperateOpRemovePass", new ShapeOperateOpRemovePass); | ||||
(void)original_graph_passes.AddPass("PrepareOptimize::ReplaceTransShapePass", new ReplaceTransShapePass); | (void)original_graph_passes.AddPass("PrepareOptimize::ReplaceTransShapePass", new ReplaceTransShapePass); | ||||
(void)original_graph_passes.AddPass("PrepareOptimize::MarkAgnosticPass", new MarkAgnosticPass); | |||||
} catch (std::bad_alloc &e) { | } catch (std::bad_alloc &e) { | ||||
GELOGE(INTERNAL_ERROR, "Add pass failed, bad memory allocation occurs."); | GELOGE(INTERNAL_ERROR, "Add pass failed, bad memory allocation occurs."); | ||||
return INTERNAL_ERROR; | return INTERNAL_ERROR; | ||||
@@ -183,6 +183,11 @@ Status AippOp::InsertAippToGraph(ComputeGraphPtr &graph, std::string &aippConfig | |||||
GE_CHECK_NOTNULL(graph); | GE_CHECK_NOTNULL(graph); | ||||
NodePtr target_input = nullptr; | NodePtr target_input = nullptr; | ||||
std::vector<std::pair<OutDataAnchorPtr, InDataAnchorPtr>> target_edges; | std::vector<std::pair<OutDataAnchorPtr, InDataAnchorPtr>> target_edges; | ||||
if (this->ConvertRelatedInputNameToRank() != SUCCESS) { | |||||
GELOGE(FAILED, "AippOp: convert related input name to rank failed."); | |||||
return FAILED; | |||||
} | |||||
GE_CHK_STATUS_RET(this->GetTargetPosition(graph, target_input, target_edges), "Get data nodes position failed"); | GE_CHK_STATUS_RET(this->GetTargetPosition(graph, target_input, target_edges), "Get data nodes position failed"); | ||||
std::map<OutDataAnchorPtr, NodePtr> out_anchors_to_aipp; | std::map<OutDataAnchorPtr, NodePtr> out_anchors_to_aipp; | ||||
@@ -410,6 +415,38 @@ Status AippOp::GetStaticTargetNode(const ComputeGraphPtr &graph, NodePtr &data_n | |||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
Status AippOp::ConvertRelatedInputNameToRank() { | |||||
GE_CHECK_NOTNULL(aipp_params_); | |||||
string related_input_name = aipp_params_->related_input_name(); | |||||
if(related_input_name.empty()) { | |||||
return SUCCESS; | |||||
} | |||||
std::vector<std::string> data_top_names = domi::GetContext().data_top_names; | |||||
GELOGI("Convert name to rank start: data size[%zu]", data_top_names.size()); | |||||
uint32_t index = 0; | |||||
bool convert_flag = false; | |||||
for (const auto &data_top_name : data_top_names) { | |||||
if (related_input_name == data_top_name) { | |||||
aipp_params_->set_related_input_rank(index); | |||||
convert_flag = true; | |||||
GELOGI("AippOp: rank: %u, top name: %s.", index, data_top_name.c_str()); | |||||
break; | |||||
} | |||||
index++; | |||||
} | |||||
if (!convert_flag) { | |||||
string error_msg = "Top name " + related_input_name + "convert rank failed, Please" | |||||
" ensure top name in aipp config is the top name of data node."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {error_msg}); | |||||
GELOGE(PARAM_INVALID, "Top name[%s] converts rank failed.", related_input_name.c_str()); | |||||
return PARAM_INVALID; | |||||
} | |||||
return SUCCESS; | |||||
} | |||||
Status AippOp::GetTargetPosition(ComputeGraphPtr graph, NodePtr &target_input, | Status AippOp::GetTargetPosition(ComputeGraphPtr graph, NodePtr &target_input, | ||||
std::vector<std::pair<OutDataAnchorPtr, InDataAnchorPtr>> &target_edges) { | std::vector<std::pair<OutDataAnchorPtr, InDataAnchorPtr>> &target_edges) { | ||||
@@ -82,6 +82,7 @@ class AippOp : public InsertOpBase { | |||||
Status AddNodeToGraph(const NodePtr &aipp_node, int64_t max_dynamic_aipp_size); | Status AddNodeToGraph(const NodePtr &aipp_node, int64_t max_dynamic_aipp_size); | ||||
Status AddAippAttrbutes(const OpDescPtr &op_desc, const std::string &aipp_cfg_path, const uint32_t &index); | Status AddAippAttrbutes(const OpDescPtr &op_desc, const std::string &aipp_cfg_path, const uint32_t &index); | ||||
Status AddAttrToAippData(const OpDescPtr &aipp_data_op_desc); | Status AddAttrToAippData(const OpDescPtr &aipp_data_op_desc); | ||||
Status ConvertRelatedInputNameToRank(); | |||||
domi::AippOpParams *aipp_params_ = nullptr; | domi::AippOpParams *aipp_params_ = nullptr; | ||||
ge::NodePtr aipp_node_ = nullptr; | ge::NodePtr aipp_node_ = nullptr; | ||||
@@ -34,6 +34,7 @@ | |||||
#include "graph/utils/op_desc_utils.h" | #include "graph/utils/op_desc_utils.h" | ||||
#include "graph/utils/tensor_utils.h" | #include "graph/utils/tensor_utils.h" | ||||
#include "graph/utils/type_utils.h" | #include "graph/utils/type_utils.h" | ||||
#include "util_insert_aipp_op.h" | |||||
using domi::AippOpParams; | using domi::AippOpParams; | ||||
@@ -115,22 +116,94 @@ void InsertNewOpUtil::ClearNewOps() { | |||||
} | } | ||||
} | } | ||||
Status InsertNewOpUtil::CheckPositionNotRepeat() { | |||||
Status InsertNewOpUtil::CheckInputNamePositionNotRepeat() { | |||||
for (int i = 0; i < insert_op_conf_->aipp_op_size(); i++) { | |||||
const domi::AippOpParams *item = insert_op_conf_->mutable_aipp_op(i); | |||||
GE_CHECK_NOTNULL(item); | |||||
for (int j = i + 1; j < insert_op_conf_->aipp_op_size(); j++) { | |||||
const domi::AippOpParams *another_item = insert_op_conf_->mutable_aipp_op(j); | |||||
GE_CHECK_NOTNULL(another_item); | |||||
if (another_item->related_input_name().empty()) { | |||||
string error_msg = "Can not both set related_input_name and related_input_rank!" | |||||
" Please ensure param is the same with the first aipp config(related_input_name)."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {error_msg}); | |||||
GELOGE(PARAM_INVALID, | |||||
"Can not both set related_input_rank and related_input_name!" | |||||
" Please ensure param is the same with the first aipp config(related_input_name)."); | |||||
return PARAM_INVALID; | |||||
} | |||||
if (item->related_input_name() == another_item->related_input_name()) { | |||||
string error_msg = "Can not insert aipp to the same postion! Please ensure related_input_name" | |||||
" param is different in different aipp config."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {error_msg}); | |||||
GELOGE(PARAM_INVALID, | |||||
"Can not insert aipp op to the same postion! Please ensure related_input_rank param " | |||||
"is different in different aipp config."); | |||||
return PARAM_INVALID; | |||||
} | |||||
} | |||||
} | |||||
return SUCCESS; | |||||
} | |||||
Status InsertNewOpUtil::CheckInputRankPositionNoRepeat() { | |||||
for (int i = 0; i < insert_op_conf_->aipp_op_size(); i++) { | for (int i = 0; i < insert_op_conf_->aipp_op_size(); i++) { | ||||
const domi::AippOpParams *item = insert_op_conf_->mutable_aipp_op(i); | const domi::AippOpParams *item = insert_op_conf_->mutable_aipp_op(i); | ||||
GE_CHECK_NOTNULL(item); | |||||
for (int j = i + 1; j < insert_op_conf_->aipp_op_size(); j++) { | for (int j = i + 1; j < insert_op_conf_->aipp_op_size(); j++) { | ||||
const domi::AippOpParams *another_item = insert_op_conf_->mutable_aipp_op(j); | const domi::AippOpParams *another_item = insert_op_conf_->mutable_aipp_op(j); | ||||
GE_IF_BOOL_EXEC(item->related_input_rank() == another_item->related_input_rank(), | |||||
string errormsg = "Can not insert aipp to the same postion! Please ensure related_input_rank" | |||||
" param is different in different aipp config."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {errormsg}); | |||||
GELOGE(PARAM_INVALID, | |||||
"Can not insert aipp op to the same postion! Please ensure related_input_rank param " | |||||
"is different in different aipp config."); | |||||
return PARAM_INVALID;); | |||||
GE_CHECK_NOTNULL(another_item); | |||||
if (!another_item->related_input_name().empty()) { | |||||
string error_msg = "Can not both set related_input_rank and related_input_name!" | |||||
" Please ensure param is the same with the first aipp config(related_input_rank)."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {error_msg}); | |||||
GELOGE(PARAM_INVALID, | |||||
"Can not both set related_input_rank and related_input_name!" | |||||
" Please ensure param is the same with the first aipp config(related_input_rank)."); | |||||
return PARAM_INVALID; | |||||
} | |||||
if (item->related_input_rank() == another_item->related_input_rank()) { | |||||
string error_msg = "Can not insert aipp to the same postion! Please ensure related_input_rank" | |||||
" param is different in different aipp config."; | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10043", {"reason"}, {error_msg}); | |||||
GELOGE(PARAM_INVALID, | |||||
"Can not insert aipp op to the same postion! Please ensure related_input_rank param " | |||||
"is different in different aipp config."); | |||||
return PARAM_INVALID; | |||||
} | |||||
} | } | ||||
} | } | ||||
return SUCCESS; | |||||
} | |||||
Status InsertNewOpUtil::CheckPositionNotRepeat() { | |||||
GE_CHECK_NOTNULL(insert_op_conf_); | |||||
if (insert_op_conf_->aipp_op_size() <= 1) { | |||||
GELOGI("Aipp op size[%d] less than 2, no need to check position repeat.", insert_op_conf_->aipp_op_size()); | |||||
return SUCCESS; | |||||
} | |||||
const domi::AippOpParams *item = insert_op_conf_->mutable_aipp_op(0); | |||||
GE_CHECK_NOTNULL(item); | |||||
string related_input_name = item->related_input_name(); | |||||
Status ret = FAILED; | |||||
if (related_input_name.empty()) { | |||||
ret = CheckInputRankPositionNoRepeat(); | |||||
} else { | |||||
ret = CheckInputNamePositionNotRepeat(); | |||||
} | |||||
if (ret != SUCCESS) { | |||||
GELOGE(FAILED, "Check position not repeat failed."); | |||||
return FAILED; | |||||
} | |||||
return SUCCESS; | return SUCCESS; | ||||
} | } | ||||
@@ -51,6 +51,10 @@ class InsertNewOpUtil { | |||||
Status GetAippParams(const std::unique_ptr<domi::AippOpParams> &aippParams, const ge::NodePtr &aipp_node); | Status GetAippParams(const std::unique_ptr<domi::AippOpParams> &aippParams, const ge::NodePtr &aipp_node); | ||||
Status CheckInputNamePositionNotRepeat(); | |||||
Status CheckInputRankPositionNoRepeat(); | |||||
Status CheckGraph(const ge::ComputeGraphPtr &graph); | Status CheckGraph(const ge::ComputeGraphPtr &graph); | ||||
InsertNewOpUtil() = default; | InsertNewOpUtil() = default; | ||||
@@ -28,7 +28,6 @@ target_include_directories(host_cpu_engine PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -36,6 +35,8 @@ target_include_directories(host_cpu_engine PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(host_cpu_engine PRIVATE | target_link_libraries(host_cpu_engine PRIVATE | ||||
@@ -67,7 +68,6 @@ target_include_directories(atc_host_cpu_engine PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -75,6 +75,8 @@ target_include_directories(atc_host_cpu_engine PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(atc_host_cpu_engine PRIVATE | target_link_libraries(atc_host_cpu_engine PRIVATE | ||||
@@ -107,7 +109,6 @@ target_include_directories(host_cpu_opskernel_builder PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -115,6 +116,8 @@ target_include_directories(host_cpu_opskernel_builder PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(host_cpu_opskernel_builder PRIVATE | target_link_libraries(host_cpu_opskernel_builder PRIVATE | ||||
@@ -141,7 +144,6 @@ target_include_directories(atc_host_cpu_opskernel_builder PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -149,6 +151,8 @@ target_include_directories(atc_host_cpu_opskernel_builder PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(atc_host_cpu_opskernel_builder PRIVATE | target_link_libraries(atc_host_cpu_opskernel_builder PRIVATE | ||||
@@ -180,7 +184,6 @@ target_include_directories(host_cpu_opskernel_builder_static PRIVATE | |||||
${GE_CODE_DIR}/inc | ${GE_CODE_DIR}/inc | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${METADEF_DIR}/inc/external/graph | ${METADEF_DIR}/inc/external/graph | ||||
@@ -188,6 +191,8 @@ target_include_directories(host_cpu_opskernel_builder_static PRIVATE | |||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(host_cpu_opskernel_builder_static PRIVATE | target_link_libraries(host_cpu_opskernel_builder_static PRIVATE | ||||
@@ -45,11 +45,12 @@ target_include_directories(atc PRIVATE | |||||
${PARSER_DIR} | ${PARSER_DIR} | ||||
${CMAKE_BINARY_DIR} | ${CMAKE_BINARY_DIR} | ||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc/toolchain | |||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
${GE_CODE_DIR}/../inc/common | ${GE_CODE_DIR}/../inc/common | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc/toolchain | |||||
) | ) | ||||
target_link_libraries(atc PRIVATE | target_link_libraries(atc PRIVATE | ||||
@@ -194,6 +194,11 @@ DEFINE_int32(op_debug_level, 0, "Optional; configure debug level of compiler. 0( | |||||
"1: open TBE compiler, export ccec file and TBE instruction mapping file; 2: open ccec compiler"); | "1: open TBE compiler, export ccec file and TBE instruction mapping file; 2: open ccec compiler"); | ||||
DEFINE_string(enable_scope_fusion_passes, "", "Optional; validate the non-general scope fusion pass," | DEFINE_string(enable_scope_fusion_passes, "", "Optional; validate the non-general scope fusion pass," | ||||
"multiple names can be set and separated by ','."); | "multiple names can be set and separated by ','."); | ||||
DEFINE_string(debug_dir, "", "Optional; the path to save the intermediate files of operator compilation"); | |||||
DEFINE_string(op_compiler_cache_dir, "", "Optional; the path to cache operator compilation files"); | |||||
DEFINE_string(op_compiler_cache_mode, "", "Optional; choose the operator compiler cache mode"); | |||||
class GFlagUtils { | class GFlagUtils { | ||||
public: | public: | ||||
@@ -999,6 +1004,9 @@ static void SetEnvForSingleOp(std::map<string, string> &options) { | |||||
options.emplace(ge::AUTO_TUNE_MODE, FLAGS_auto_tune_mode); | options.emplace(ge::AUTO_TUNE_MODE, FLAGS_auto_tune_mode); | ||||
options.emplace(ge::GRAPH_MEMORY_MAX_SIZE, kGraphMemoryManagerMallocMaxSize); | options.emplace(ge::GRAPH_MEMORY_MAX_SIZE, kGraphMemoryManagerMallocMaxSize); | ||||
options.emplace(ge::OP_DEBUG_LEVEL, to_string(FLAGS_op_debug_level)); | options.emplace(ge::OP_DEBUG_LEVEL, to_string(FLAGS_op_debug_level)); | ||||
options.emplace(ge::DEBUG_DIR, FLAGS_debug_dir); | |||||
options.emplace(ge::OP_COMPILER_CACHE_DIR, FLAGS_op_compiler_cache_dir); | |||||
options.emplace(ge::OP_COMPILER_CACHE_MODE, FLAGS_op_compiler_cache_mode); | |||||
} | } | ||||
domi::Status GenerateSingleOp(const std::string& json_file_path) { | domi::Status GenerateSingleOp(const std::string& json_file_path) { | ||||
@@ -1131,6 +1139,12 @@ domi::Status GenerateOmModel() { | |||||
options.insert(std::pair<string, string>(string(ge::ENABLE_SINGLE_STREAM), FLAGS_enable_single_stream)); | options.insert(std::pair<string, string>(string(ge::ENABLE_SINGLE_STREAM), FLAGS_enable_single_stream)); | ||||
options.insert(std::pair<string, string>(string(ge::DEBUG_DIR), FLAGS_debug_dir)); | |||||
options.insert(std::pair<string, string>(string(ge::OP_COMPILER_CACHE_DIR), FLAGS_op_compiler_cache_dir)); | |||||
options.insert(std::pair<string, string>(string(ge::OP_COMPILER_CACHE_MODE), FLAGS_op_compiler_cache_mode)); | |||||
SetDynamicInputSizeOptions(); | SetDynamicInputSizeOptions(); | ||||
if (!FLAGS_save_original_model.empty()) { | if (!FLAGS_save_original_model.empty()) { | ||||
@@ -45,6 +45,9 @@ message AippOpParams { | |||||
// 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | // 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | ||||
uint32 related_input_rank = 2; | uint32 related_input_rank = 2; | ||||
// related_input_name is optional and the top name of data node which inserts aipp | |||||
string related_input_name = 6; | |||||
// input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | // input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | ||||
// 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | // 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | ||||
// 配置值 <= Data算子输出边的个数。 | // 配置值 <= Data算子输出边的个数。 | ||||
@@ -226,10 +226,11 @@ bool SingleOpParser::Validate(const SingleOpDesc &op_desc) { | |||||
} | } | ||||
int index = 0; | int index = 0; | ||||
for (auto &tensor_desc : op_desc.output_desc) { | |||||
for (auto &tensor_desc : op_desc.input_desc) { | |||||
if ((tensor_desc.type == DT_UNDEFINED && tensor_desc.format != FORMAT_RESERVED) || | if ((tensor_desc.type == DT_UNDEFINED && tensor_desc.format != FORMAT_RESERVED) || | ||||
(tensor_desc.type != DT_UNDEFINED && tensor_desc.format == FORMAT_RESERVED)){ | (tensor_desc.type != DT_UNDEFINED && tensor_desc.format == FORMAT_RESERVED)){ | ||||
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "index"}, {"output", std::to_string(index)}); | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "type", "index"}, | |||||
{"intput", "datatype or format", std::to_string(index)}); | |||||
GELOGE(PARAM_INVALID, "Input's dataType or format is invalid when the index is %d", index); | GELOGE(PARAM_INVALID, "Input's dataType or format is invalid when the index is %d", index); | ||||
return false; | return false; | ||||
} | } | ||||
@@ -239,13 +240,15 @@ bool SingleOpParser::Validate(const SingleOpDesc &op_desc) { | |||||
index = 0; | index = 0; | ||||
for (auto &tensor_desc : op_desc.output_desc) { | for (auto &tensor_desc : op_desc.output_desc) { | ||||
if (tensor_desc.type == DT_UNDEFINED) { | if (tensor_desc.type == DT_UNDEFINED) { | ||||
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "index"}, {"output", std::to_string(index)}); | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "type", "index"}, | |||||
{"output", "datatype", std::to_string(index)}); | |||||
GELOGE(PARAM_INVALID, "Output's dataType is invalid when the index is %d", index); | GELOGE(PARAM_INVALID, "Output's dataType is invalid when the index is %d", index); | ||||
return false; | return false; | ||||
} | } | ||||
if (tensor_desc.format == FORMAT_RESERVED) { | if (tensor_desc.format == FORMAT_RESERVED) { | ||||
ErrorManager::GetInstance().ATCReportErrMessage("E10028", {"input", "index"}, {"output", std::to_string(index)}); | |||||
ErrorManager::GetInstance().ATCReportErrMessage("E10027", {"input", "type", "index"}, | |||||
{"output", "format", std::to_string(index)}); | |||||
GELOGE(PARAM_INVALID, "Output's format is invalid when the index is %d", index); | GELOGE(PARAM_INVALID, "Output's format is invalid when the index is %d", index); | ||||
return false; | return false; | ||||
} | } | ||||
@@ -1 +1 @@ | |||||
optimizer:["aicpu_tf_optimizer","AIcoreEngine","VectorEngine","aicpu_ascend_optimizer","hccl_graph_optimizer", "hvd_graph_optimizer", "DNN_VM_RTS_GRAPH_OPTIMIZER_STORE"] | |||||
optimizer:["aicpu_tf_optimizer","aicpu_ascend_optimizer","AIcoreEngine","VectorEngine","hccl_graph_optimizer", "hvd_graph_optimizer", "DNN_VM_RTS_GRAPH_OPTIMIZER_STORE"] |
@@ -21,13 +21,14 @@ target_include_directories(engine PRIVATE | |||||
${GE_CODE_DIR}/inc/framework | ${GE_CODE_DIR}/inc/framework | ||||
${GE_CODE_DIR}/inc/framework/common | ${GE_CODE_DIR}/inc/framework/common | ||||
${GE_CODE_DIR}/inc/external | ${GE_CODE_DIR}/inc/external | ||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
${METADEF_DIR}/inc | ${METADEF_DIR}/inc | ||||
${METADEF_DIR}/inc/external | ${METADEF_DIR}/inc/external | ||||
${CMAKE_BINARY_DIR} | ${CMAKE_BINARY_DIR} | ||||
${CMAKE_BINARY_DIR}/proto/ge | ${CMAKE_BINARY_DIR}/proto/ge | ||||
#### yellow zone #### | #### yellow zone #### | ||||
${GE_CODE_DIR}/../inc | ${GE_CODE_DIR}/../inc | ||||
#### blue zone #### | |||||
${GE_CODE_DIR}/third_party/fwkacllib/inc | |||||
) | ) | ||||
target_link_libraries(engine PRIVATE | target_link_libraries(engine PRIVATE | ||||
@@ -45,6 +45,9 @@ message AippOpParams { | |||||
// 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | // 标识对模型的第几个输入做AIPP处理,例如模型有两个输入,需要对第2个输入做AIPP,则配置related_input_rank为1。 | ||||
uint32 related_input_rank = 2; | uint32 related_input_rank = 2; | ||||
// related_input_name is optional and the top name of data node which inserts aipp | |||||
string related_input_name = 6; | |||||
// input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | // input_edge_idx参数为可选,类型为整型,配置范围为>=0。 | ||||
// 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | // 配置该参数的作用,在于对Data算子不同的输出做不同的AIPP处理,如果该参数没有配置,默认对related_input_rank指定的模型输入的所有输出边做AIPP。 | ||||
// 配置值 <= Data算子输出边的个数。 | // 配置值 <= Data算子输出边的个数。 | ||||
@@ -1038,6 +1038,7 @@ void UpdateOmgCtxWithParserCtx() { | |||||
domi::GetContext().out_top_names = GetParserContext().out_top_names; | domi::GetContext().out_top_names = GetParserContext().out_top_names; | ||||
domi::GetContext().user_out_nodes_top_vec = GetParserContext().user_out_nodes_top_vec; | domi::GetContext().user_out_nodes_top_vec = GetParserContext().user_out_nodes_top_vec; | ||||
domi::GetContext().default_out_nodes = GetParserContext().default_out_nodes; | domi::GetContext().default_out_nodes = GetParserContext().default_out_nodes; | ||||
domi::GetContext().data_top_names = GetParserContext().data_top_names; | |||||
} | } | ||||
void UpdateParserCtxWithOmgCtx() { | void UpdateParserCtxWithOmgCtx() { | ||||
@@ -1054,5 +1055,6 @@ void UpdateParserCtxWithOmgCtx() { | |||||
GetParserContext().input_nodes_format_map = domi::GetContext().input_nodes_format_map; | GetParserContext().input_nodes_format_map = domi::GetContext().input_nodes_format_map; | ||||
GetParserContext().out_top_names = domi::GetContext().out_top_names; | GetParserContext().out_top_names = domi::GetContext().out_top_names; | ||||
GetParserContext().user_out_nodes_top_vec = domi::GetContext().user_out_nodes_top_vec; | GetParserContext().user_out_nodes_top_vec = domi::GetContext().user_out_nodes_top_vec; | ||||
GetParserContext().data_top_names = domi::GetContext().data_top_names; | |||||
} | } | ||||
} // namespace ge | } // namespace ge |
@@ -190,6 +190,7 @@ Status SingleOpModel::LoadAllNodes() { | |||||
} | } | ||||
ge_model->GetTBEKernelStore().LoadTBEKernelBinToOpDesc(op_desc); | ge_model->GetTBEKernelStore().LoadTBEKernelBinToOpDesc(op_desc); | ||||
ge_model->GetCustAICPUKernelStore().LoadCustAICPUKernelBinToOpDesc(op_desc); | |||||
} | } | ||||
return SUCCESS; | return SUCCESS; | ||||
@@ -247,7 +248,7 @@ Status SingleOpModel::BuildTaskList(SingleOp &single_op) { | |||||
single_op.arg_table_.resize(single_op.input_sizes_.size() + single_op.output_sizes_.size()); | single_op.arg_table_.resize(single_op.input_sizes_.size() + single_op.output_sizes_.size()); | ||||
ParseArgTable(tbe_task, single_op); | ParseArgTable(tbe_task, single_op); | ||||
single_op.tasks_.emplace_back(tbe_task); | single_op.tasks_.emplace_back(tbe_task); | ||||
} else if (kernel_type == cce::ccKernelType::AI_CPU) { | |||||
} else if (kernel_type == cce::ccKernelType::AI_CPU || kernel_type == cce::ccKernelType::CUST_AI_CPU) { | |||||
GELOGD("Building AICPU_CC task"); | GELOGD("Building AICPU_CC task"); | ||||
OpTask *task = nullptr; | OpTask *task = nullptr; | ||||
auto ret = BuildCpuKernelTask(task_def.kernel(), &task); | auto ret = BuildCpuKernelTask(task_def.kernel(), &task); | ||||
@@ -256,7 +257,7 @@ Status SingleOpModel::BuildTaskList(SingleOp &single_op) { | |||||
} | } | ||||
single_op.tasks_.emplace_back(task); | single_op.tasks_.emplace_back(task); | ||||
} else { | } else { | ||||
GELOGE(UNSUPPORTED, "Only TBE kernel and AI_CPU kernel are supported, but got %u", context.kernel_type()); | |||||
GELOGE(UNSUPPORTED, "Only TBE, AI_CPU, CUST_AI_CPU kernel are supported, but got %u", context.kernel_type()); | |||||
return UNSUPPORTED; | return UNSUPPORTED; | ||||
} | } | ||||
} else if (task_type == RT_MODEL_TASK_KERNEL_EX) { | } else if (task_type == RT_MODEL_TASK_KERNEL_EX) { | ||||
@@ -391,13 +392,13 @@ Status SingleOpModel::BuildModelTaskKernel(const TaskDef &task_def, DynamicSingl | |||||
TbeOpTask *tbe_task = nullptr; | TbeOpTask *tbe_task = nullptr; | ||||
GE_CHK_STATUS_RET_NOLOG(BuildKernelTask(task_def.kernel(), &tbe_task)); | GE_CHK_STATUS_RET_NOLOG(BuildKernelTask(task_def.kernel(), &tbe_task)); | ||||
single_op.op_task_.reset(tbe_task); | single_op.op_task_.reset(tbe_task); | ||||
} else if (kernel_type == cce::ccKernelType::AI_CPU) { | |||||
} else if (kernel_type == cce::ccKernelType::AI_CPU || kernel_type == cce::ccKernelType::CUST_AI_CPU) { | |||||
GELOGD("Building AICPU_CC task"); | GELOGD("Building AICPU_CC task"); | ||||
OpTask *task = nullptr; | OpTask *task = nullptr; | ||||
GE_CHK_STATUS_RET_NOLOG(BuildCpuKernelTask(task_def.kernel(), &task)); | GE_CHK_STATUS_RET_NOLOG(BuildCpuKernelTask(task_def.kernel(), &task)); | ||||
single_op.op_task_.reset(task); | single_op.op_task_.reset(task); | ||||
} else { | } else { | ||||
GELOGE(UNSUPPORTED, "Only TBE kernel and AI_CPU kernel are supported, but got %u", context.kernel_type()); | |||||
GELOGE(UNSUPPORTED, "Only TBE, AI_CPU, CUST_AI_CPU kernel are supported, but got %u", context.kernel_type()); | |||||
return UNSUPPORTED; | return UNSUPPORTED; | ||||
} | } | ||||
return SUCCESS; | return SUCCESS; | ||||
@@ -15,6 +15,8 @@ | |||||
*/ | */ | ||||
#include "single_op/task/aicpu_kernel_task_builder.h" | #include "single_op/task/aicpu_kernel_task_builder.h" | ||||
#include "cce/taskdown_common.hpp" | |||||
#include "graph/load/new_model_manager/model_manager.h" | |||||
namespace ge { | namespace ge { | ||||
AiCpuCCTaskBuilder::AiCpuCCTaskBuilder(const OpDescPtr &op_desc, const domi::KernelDef &kernel_def) | AiCpuCCTaskBuilder::AiCpuCCTaskBuilder(const OpDescPtr &op_desc, const domi::KernelDef &kernel_def) | ||||
@@ -55,6 +57,14 @@ Status AiCpuCCTaskBuilder::BuildTask(AiCpuCCTask &task) { | |||||
task.SetkernelName(kernel_name); | task.SetkernelName(kernel_name); | ||||
task.op_desc_ = op_desc_; | task.op_desc_ = op_desc_; | ||||
const auto &context = kernel_def_.context(); | |||||
auto kernel_type = static_cast<cce::ccKernelType>(context.kernel_type()); | |||||
if (kernel_type == cce::ccKernelType::CUST_AI_CPU) { | |||||
task.is_custom_ = true; | |||||
task.dump_flag_ |= RT_KERNEL_CUSTOM_AICPU; | |||||
GE_CHK_STATUS_RET(ModelManager::GetInstance()->LoadCustAicpuSo(op_desc_, so_name), "launch cust aicpu so failed"); | |||||
} | |||||
task.num_inputs_ = op_desc_->GetInputsSize(); | task.num_inputs_ = op_desc_->GetInputsSize(); | ||||
task.num_outputs_ = op_desc_->GetOutputsSize(); | task.num_outputs_ = op_desc_->GetOutputsSize(); | ||||
@@ -45,6 +45,7 @@ std::vector<std::vector<void *>> BuildTaskUtils::GetAddresses(const OpDescPtr &o | |||||
runtime_para.logic_var_base = kLogicVarBase; | runtime_para.logic_var_base = kLogicVarBase; | ||||
runtime_para.var_base = kVarBase; | runtime_para.var_base = kVarBase; | ||||
runtime_para.session_id = kSessionId; | runtime_para.session_id = kSessionId; | ||||
runtime_para.is_single_op = true; | |||||
ret.emplace_back(ModelUtils::GetInputDataAddrs(runtime_para, op_desc)); | ret.emplace_back(ModelUtils::GetInputDataAddrs(runtime_para, op_desc)); | ||||
ret.emplace_back(ModelUtils::GetOutputDataAddrs(runtime_para, op_desc)); | ret.emplace_back(ModelUtils::GetOutputDataAddrs(runtime_para, op_desc)); | ||||
@@ -280,8 +280,6 @@ Status AiCpuBaseTask::UpdateExtInfo(const std::vector<GeTensorDesc> &input_desc, | |||||
for (size_t j = 0; j < num_outputs_; ++j) { | for (size_t j = 0; j < num_outputs_; ++j) { | ||||
GE_CHK_STATUS_RET(aicpu_ext_handle_->UpdateOutputShapeAndType(j, output_desc[j]), | GE_CHK_STATUS_RET(aicpu_ext_handle_->UpdateOutputShapeAndType(j, output_desc[j]), | ||||
"Output[%zu] UpdateOutputShapeAndType failed.", j); | "Output[%zu] UpdateOutputShapeAndType failed.", j); | ||||
// debug code | |||||
GELOGD("No input and output, no need update ext info."); | |||||
} | } | ||||
} | } | ||||
@@ -669,9 +667,10 @@ Status AiCpuCCTask::LaunchKernel(rtStream_t stream) { | |||||
kernel_name_.data()); | kernel_name_.data()); | ||||
// sm_desc is nullptr, because l2 buffer does not support | // sm_desc is nullptr, because l2 buffer does not support | ||||
auto *sm_desc = reinterpret_cast<rtSmDesc_t *>(sm_desc_); | auto *sm_desc = reinterpret_cast<rtSmDesc_t *>(sm_desc_); | ||||
auto ret = | |||||
rtCpuKernelLaunch(static_cast<const void *>(so_name_.data()), static_cast<const void *>(kernel_name_.data()), | |||||
block_dim_, args_.get(), static_cast<uint32_t>(arg_size_), sm_desc, stream); | |||||
auto ret = rtCpuKernelLaunchWithFlag(static_cast<const void *>(so_name_.data()), | |||||
static_cast<const void *>(kernel_name_.data()), | |||||
block_dim_, args_.get(), static_cast<uint32_t>(arg_size_), | |||||
sm_desc, stream, dump_flag_); | |||||
if (ret != RT_ERROR_NONE) { | if (ret != RT_ERROR_NONE) { | ||||
GELOGE(RT_FAILED, "Invoke rtCpuKernelLaunch failed. ret = %d", ret); | GELOGE(RT_FAILED, "Invoke rtCpuKernelLaunch failed. ret = %d", ret); | ||||
return RT_FAILED; | return RT_FAILED; | ||||
@@ -234,6 +234,8 @@ private: | |||||
uint32_t block_dim_ = 1; | uint32_t block_dim_ = 1; | ||||
void *sm_desc_ = nullptr; | void *sm_desc_ = nullptr; | ||||
void *io_addr_ = nullptr; | void *io_addr_ = nullptr; | ||||
bool is_custom_ = false; | |||||
uint32_t dump_flag_ = RT_KERNEL_DEFAULT; | |||||
}; | }; | ||||
} // namespace ge | } // namespace ge | ||||
@@ -222,6 +222,18 @@ const char *const OPTION_GE_MAX_DUMP_OP_NUM = "ge.maxDumpOpNum"; | |||||
// Its value should be "0" or "1", default value is "1" | // Its value should be "0" or "1", default value is "1" | ||||
const char *const ENABLE_PRINT_OP_PASS = "ge.enablePrintOpPass"; | const char *const ENABLE_PRINT_OP_PASS = "ge.enablePrintOpPass"; | ||||
// Configure operator compilation path | |||||
// Its value should be file path, default value is "./" | |||||
const char *const DEBUG_DIR = "ge.debugDir"; | |||||
// Configure operator compiler cache path | |||||
// Its value should be file path, default value is "./" | |||||
const char *const OP_COMPILER_CACHE_DIR = "ge.op_compiler_cache_dir"; | |||||
// Configure operator compiler cache mode | |||||
// Its value should be "disable", "enable" or "force", default value is "disable" | |||||
const char *const OP_COMPILER_CACHE_MODE = "ge.op_compiler_cache_mode"; | |||||
// Configure whether to use single stream. | // Configure whether to use single stream. | ||||
// Its value should be "true" or "false", default value is "false" | // Its value should be "true" or "false", default value is "false" | ||||
const char *const ENABLE_SINGLE_STREAM = "ge.enableSingleStream"; | const char *const ENABLE_SINGLE_STREAM = "ge.enableSingleStream"; | ||||
@@ -238,10 +250,10 @@ enum GraphRunMode { PREDICTION = 0, TRAIN }; | |||||
// Input/Output tensor info | // Input/Output tensor info | ||||
struct InputTensorInfo { | struct InputTensorInfo { | ||||
uint32_t data_type; // data type | |||||
std::vector<int64_t> dims; // shape description | |||||
void *data; // tensor data | |||||
int64_t length; // tensor length | |||||
uint32_t data_type; // data type | |||||
std::vector<int64_t> dims; // shape description | |||||
void *data; // tensor data | |||||
int64_t length; // tensor length | |||||
}; | }; | ||||
struct OutputTensorInfo { | struct OutputTensorInfo { | ||||
@@ -250,11 +262,8 @@ struct OutputTensorInfo { | |||||
std::unique_ptr<uint8_t[]> data; // tensor data | std::unique_ptr<uint8_t[]> data; // tensor data | ||||
int64_t length; // tensor length | int64_t length; // tensor length | ||||
OutputTensorInfo() : data_type(0), dims({}), data(nullptr), length(0) {} | OutputTensorInfo() : data_type(0), dims({}), data(nullptr), length(0) {} | ||||
OutputTensorInfo(OutputTensorInfo &&out) : | |||||
data_type(out.data_type), | |||||
dims(out.dims), | |||||
data(std::move(out.data)), | |||||
length(out.length) {} | |||||
OutputTensorInfo(OutputTensorInfo &&out) | |||||
: data_type(out.data_type), dims(out.dims), data(std::move(out.data)), length(out.length) {} | |||||
OutputTensorInfo &operator=(OutputTensorInfo &&out) { | OutputTensorInfo &operator=(OutputTensorInfo &&out) { | ||||
if (this != &out) { | if (this != &out) { | ||||
@@ -273,67 +282,60 @@ using Status = uint32_t; | |||||
using RunAsyncCallback = std::function<void(Status, std::vector<ge::OutputTensorInfo> &)>; | using RunAsyncCallback = std::function<void(Status, std::vector<ge::OutputTensorInfo> &)>; | ||||
// for ir build | // for ir build | ||||
namespace ir_option { | namespace ir_option { | ||||
static const char *const INPUT_FORMAT = "input_format"; | |||||
static const char *const INPUT_SHAPE = "input_shape"; | |||||
static const char *const OP_NAME_MAP = "op_name_map"; | |||||
static const char *const DYNAMIC_BATCH_SIZE = kDynamicBatchSize; | |||||
static const char *const DYNAMIC_IMAGE_SIZE = kDynamicImageSize; | |||||
static const char *const DYNAMIC_DIMS = kDynamicDims; | |||||
static const char *const INSERT_OP_FILE = ge::INSERT_OP_FILE.c_str(); | |||||
static const char *const PRECISION_MODE = ge::PRECISION_MODE.c_str(); | |||||
static const char *const EXEC_DISABLE_REUSED_MEMORY = ge::OPTION_EXEC_DISABLE_REUSED_MEMORY; | |||||
static const char *const AUTO_TUNE_MODE = ge::AUTO_TUNE_MODE.c_str(); | |||||
static const char *const CORE_TYPE = ge::CORE_TYPE.c_str(); | |||||
static const char *const SOC_VERSION = ge::SOC_VERSION.c_str(); | |||||
static const char *const ENABLE_SINGLE_STREAM = ge::ENABLE_SINGLE_STREAM; | |||||
static const char *const AICORE_NUM = ge::AICORE_NUM.c_str(); | |||||
static const char *const FUSION_SWITCH_FILE = ge::FUSION_SWITCH_FILE.c_str(); | |||||
static const char *const ENABLE_SMALL_CHANNEL = ge::ENABLE_SMALL_CHANNEL.c_str(); | |||||
static const char *const OP_SELECT_IMPL_MODE = ge::OP_SELECT_IMPL_MODE.c_str(); | |||||
static const char *const OUTPUT_TYPE = ge::OUTPUT_DATATYPE.c_str(); | |||||
static const char *const BUFFER_OPTIMIZE = ge::BUFFER_OPTIMIZE.c_str(); | |||||
static const char *const ENABLE_COMPRESS_WEIGHT = ge::ENABLE_COMPRESS_WEIGHT.c_str(); | |||||
static const char *const COMPRESS_WEIGHT_CONF = "compress_weight_conf"; | |||||
static const char *const OUT_NODES = ge::OUTPUT_NODE_NAME.c_str(); | |||||
static const char *const INPUT_FP16_NODES = ge::INPUT_FP16_NODES.c_str(); | |||||
static const char *const LOG_LEVEL = "log"; | |||||
static const char *const OPTYPELIST_FOR_IMPLMODE = ge::OPTYPELIST_FOR_IMPLMODE.c_str(); | |||||
// for interface: aclgrphBuildModel | |||||
const std::set<std::string> ir_builder_suppported_options = { | |||||
INPUT_FORMAT, | |||||
INPUT_SHAPE, | |||||
OP_NAME_MAP, | |||||
DYNAMIC_BATCH_SIZE, | |||||
DYNAMIC_IMAGE_SIZE, | |||||
DYNAMIC_DIMS, | |||||
INSERT_OP_FILE, | |||||
PRECISION_MODE, | |||||
EXEC_DISABLE_REUSED_MEMORY, | |||||
AUTO_TUNE_MODE, | |||||
OUTPUT_TYPE, | |||||
OUT_NODES, | |||||
INPUT_FP16_NODES, | |||||
LOG_LEVEL | |||||
}; | |||||
// for interface: aclgrphBuildInitialize | |||||
const std::set<std::string> global_options = { | |||||
CORE_TYPE, | |||||
SOC_VERSION, | |||||
BUFFER_OPTIMIZE, | |||||
ENABLE_COMPRESS_WEIGHT, | |||||
COMPRESS_WEIGHT_CONF, | |||||
PRECISION_MODE, | |||||
EXEC_DISABLE_REUSED_MEMORY, | |||||
AUTO_TUNE_MODE, | |||||
ENABLE_SINGLE_STREAM, | |||||
AICORE_NUM, | |||||
FUSION_SWITCH_FILE, | |||||
ENABLE_SMALL_CHANNEL, | |||||
OP_SELECT_IMPL_MODE, | |||||
OPTYPELIST_FOR_IMPLMODE | |||||
}; | |||||
} | |||||
static const char *const INPUT_FORMAT = "input_format"; | |||||
static const char *const INPUT_SHAPE = "input_shape"; | |||||
static const char *const OP_NAME_MAP = "op_name_map"; | |||||
static const char *const DYNAMIC_BATCH_SIZE = kDynamicBatchSize; | |||||
static const char *const DYNAMIC_IMAGE_SIZE = kDynamicImageSize; | |||||
static const char *const DYNAMIC_DIMS = kDynamicDims; | |||||
static const char *const INSERT_OP_FILE = ge::INSERT_OP_FILE.c_str(); | |||||
static const char *const PRECISION_MODE = ge::PRECISION_MODE.c_str(); | |||||
static const char *const EXEC_DISABLE_REUSED_MEMORY = ge::OPTION_EXEC_DISABLE_REUSED_MEMORY; | |||||
static const char *const AUTO_TUNE_MODE = ge::AUTO_TUNE_MODE.c_str(); | |||||
static const char *const CORE_TYPE = ge::CORE_TYPE.c_str(); | |||||
static const char *const SOC_VERSION = ge::SOC_VERSION.c_str(); | |||||
static const char *const ENABLE_SINGLE_STREAM = ge::ENABLE_SINGLE_STREAM; | |||||
static const char *const AICORE_NUM = ge::AICORE_NUM.c_str(); | |||||
static const char *const FUSION_SWITCH_FILE = ge::FUSION_SWITCH_FILE.c_str(); | |||||
static const char *const ENABLE_SMALL_CHANNEL = ge::ENABLE_SMALL_CHANNEL.c_str(); | |||||
static const char *const OP_SELECT_IMPL_MODE = ge::OP_SELECT_IMPL_MODE.c_str(); | |||||
static const char *const OUTPUT_TYPE = ge::OUTPUT_DATATYPE.c_str(); | |||||
static const char *const BUFFER_OPTIMIZE = ge::BUFFER_OPTIMIZE.c_str(); | |||||
static const char *const ENABLE_COMPRESS_WEIGHT = ge::ENABLE_COMPRESS_WEIGHT.c_str(); | |||||
static const char *const COMPRESS_WEIGHT_CONF = "compress_weight_conf"; | |||||
static const char *const OUT_NODES = ge::OUTPUT_NODE_NAME.c_str(); | |||||
static const char *const INPUT_FP16_NODES = ge::INPUT_FP16_NODES.c_str(); | |||||
static const char *const LOG_LEVEL = "log"; | |||||
static const char *const OPTYPELIST_FOR_IMPLMODE = ge::OPTYPELIST_FOR_IMPLMODE.c_str(); | |||||
static const char *const DEBUG_DIR = ge::DEBUG_DIR; | |||||
static const char *const OP_COMPILER_CACHE_DIR = ge::OP_COMPILER_CACHE_DIR; | |||||
static const char *const OP_COMPILER_CACHE_MODE = ge::OP_COMPILER_CACHE_MODE; | |||||
// for interface: aclgrphBuildModel | |||||
const std::set<std::string> ir_builder_suppported_options = { | |||||
INPUT_FORMAT, INPUT_SHAPE, OP_NAME_MAP, | |||||
DYNAMIC_BATCH_SIZE, DYNAMIC_IMAGE_SIZE, DYNAMIC_DIMS, | |||||
INSERT_OP_FILE, PRECISION_MODE, EXEC_DISABLE_REUSED_MEMORY, | |||||
AUTO_TUNE_MODE, OUTPUT_TYPE, OUT_NODES, | |||||
INPUT_FP16_NODES, LOG_LEVEL}; | |||||
// for interface: aclgrphBuildInitialize | |||||
const std::set<std::string> global_options = {CORE_TYPE, | |||||
SOC_VERSION, | |||||
BUFFER_OPTIMIZE, | |||||
ENABLE_COMPRESS_WEIGHT, | |||||
COMPRESS_WEIGHT_CONF, | |||||
PRECISION_MODE, | |||||
EXEC_DISABLE_REUSED_MEMORY, | |||||
AUTO_TUNE_MODE, | |||||
ENABLE_SINGLE_STREAM, | |||||
AICORE_NUM, | |||||
FUSION_SWITCH_FILE, | |||||
ENABLE_SMALL_CHANNEL, | |||||
OP_SELECT_IMPL_MODE, | |||||
OPTYPELIST_FOR_IMPLMODE, | |||||
DEBUG_DIR, | |||||
OP_COMPILER_CACHE_DIR, | |||||
OP_COMPILER_CACHE_MODE}; | |||||
} // namespace ir_option | |||||
} // namespace ge | } // namespace ge | ||||
#endif // INC_EXTERNAL_GE_GE_API_TYPES_H_ | #endif // INC_EXTERNAL_GE_GE_API_TYPES_H_ |
@@ -449,6 +449,7 @@ REGISTER_OPTYPE_DECLARE(MEMCPYASYNC, "MemcpyAsync"); | |||||
REGISTER_OPTYPE_DECLARE(MEMCPYADDRASYNC, "MemcpyAddrAsync"); | REGISTER_OPTYPE_DECLARE(MEMCPYADDRASYNC, "MemcpyAddrAsync"); | ||||
REGISTER_OPTYPE_DECLARE(STREAMMERGE, "StreamMerge"); | REGISTER_OPTYPE_DECLARE(STREAMMERGE, "StreamMerge"); | ||||
REGISTER_OPTYPE_DECLARE(ENDGRAPH, "EndGraph"); | REGISTER_OPTYPE_DECLARE(ENDGRAPH, "EndGraph"); | ||||
REGISTER_OPTYPE_DECLARE(MODELEXIT, "ModelExit"); | |||||
REGISTER_OPTYPE_DECLARE(SEND, "Send"); | REGISTER_OPTYPE_DECLARE(SEND, "Send"); | ||||
REGISTER_OPTYPE_DECLARE(RECV, "Recv"); | REGISTER_OPTYPE_DECLARE(RECV, "Recv"); | ||||
REGISTER_OPTYPE_DECLARE(ENDOFSEQUENCE, "EndOfSequence"); | REGISTER_OPTYPE_DECLARE(ENDOFSEQUENCE, "EndOfSequence"); | ||||
@@ -100,6 +100,8 @@ struct OmgContext { | |||||
std::vector<std::string> net_out_nodes; | std::vector<std::string> net_out_nodes; | ||||
// net out nodes top names(only caffe has top) | // net out nodes top names(only caffe has top) | ||||
std::vector<std::string> out_top_names; | std::vector<std::string> out_top_names; | ||||
// net data nodes top names(only caffe has top) | |||||
std::vector<std::string> data_top_names; | |||||
// preferential format used by the entire network | // preferential format used by the entire network | ||||
domiTensorFormat_t net_format = DOMI_TENSOR_RESERVED; | domiTensorFormat_t net_format = DOMI_TENSOR_RESERVED; | ||||
domi::FrameworkType type = domi::FRAMEWORK_RESERVED; | domi::FrameworkType type = domi::FRAMEWORK_RESERVED; | ||||
@@ -49,6 +49,8 @@ struct ParserContext { | |||||
std::vector<std::string> user_out_nodes_top_vec; | std::vector<std::string> user_out_nodes_top_vec; | ||||
// net out nodes (where user_out_nodes or leaf nodes) | // net out nodes (where user_out_nodes or leaf nodes) | ||||
std::vector<std::string> net_out_nodes; | std::vector<std::string> net_out_nodes; | ||||
// net data nodes top names(only caffe has top) | |||||
std::vector<std::string> data_top_names; | |||||
// net out nodes top names(only caffe has top) | // net out nodes top names(only caffe has top) | ||||
std::vector<std::string> out_top_names; | std::vector<std::string> out_top_names; | ||||
// Whether to use dynamic batch size or dynamic image size | // Whether to use dynamic batch size or dynamic image size | ||||
@@ -57,9 +59,12 @@ struct ParserContext { | |||||
domi::domiTensorFormat_t format = domi::DOMI_TENSOR_ND; | domi::domiTensorFormat_t format = domi::DOMI_TENSOR_ND; | ||||
domi::FrameworkType type = domi::FRAMEWORK_RESERVED; | domi::FrameworkType type = domi::FRAMEWORK_RESERVED; | ||||
RunMode run_mode = ONLY_PRE_CHECK; | RunMode run_mode = ONLY_PRE_CHECK; | ||||
std::string custom_proto_path; // save caffe custom proto path, used by caffe parse | |||||
std::string caffe_proto_path; // save caffe proto path, used by caffe parse | |||||
std::string enable_scope_fusion_passes; // name of the pass that needs to take effect | |||||
// save caffe custom proto path, used by caffe parse | |||||
std::string custom_proto_path; | |||||
// save caffe proto path, used by caffe parse | |||||
std::string caffe_proto_path; | |||||
// name of the pass that needs to take effect | |||||
std::string enable_scope_fusion_passes; | |||||
}; | }; | ||||
ParserContext &GetParserContext(); | ParserContext &GetParserContext(); | ||||
@@ -1 +1 @@ | |||||
Subproject commit 1cc55bcae09902b3d158993dd57bfbd1d3337066 | |||||
Subproject commit cc9de48a7779cf95cab90a23db608421a691fd12 |
@@ -1 +1 @@ | |||||
Subproject commit db4e6070bb2cec01cead264a44ceae07e7f3048e | |||||
Subproject commit f86c751307835d85329260feb51b559d378b5710 |
@@ -190,6 +190,7 @@ file(GLOB_RECURSE DISTINCT_GRAPH_LOAD_SRC_FILES ${CMAKE_CURRENT_SOURCE_DIR} | |||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/stream_active_task_info.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/stream_active_task_info.cc" | ||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/stream_switch_task_info.cc" | ||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/end_graph_task_info.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/end_graph_task_info.cc" | ||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/model_exit_task_info.cc" | |||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/super_kernel/super_kernel.cc" | ||||
"${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/new_model_manager/task_info/super_kernel/super_kernel_factory.cc" | ||||
"${GE_SOURCE_DIR}/src/ge/graph/load/output/output.cc" | "${GE_SOURCE_DIR}/src/ge/graph/load/output/output.cc" | ||||
@@ -17,10 +17,10 @@ | |||||
#ifndef _MMPA_API_H_ | #ifndef _MMPA_API_H_ | ||||
#define _MMPA_API_H_ | #define _MMPA_API_H_ | ||||
#define LINUX 0 | |||||
#define WIN 1 | |||||
#define LINUX 0 | |||||
#define WIN 1 | |||||
#if(OS_TYPE == LINUX) | |||||
#if(OS_TYPE == LINUX) //lint !e553 | |||||
#ifndef _GNU_SOURCE | #ifndef _GNU_SOURCE | ||||
#define _GNU_SOURCE | #define _GNU_SOURCE | ||||
@@ -75,6 +75,7 @@ | |||||
#include <sys/wait.h> | #include <sys/wait.h> | ||||
#include <sys/statvfs.h> | #include <sys/statvfs.h> | ||||
#include <sys/prctl.h> | #include <sys/prctl.h> | ||||
#include <sys/inotify.h> | |||||
#include "securec.h" | #include "securec.h" | ||||
@@ -84,7 +85,7 @@ | |||||
#endif | #endif | ||||
#if(OS_TYPE == WIN) | |||||
#if(OS_TYPE == WIN) //lint !e553 | |||||
#include <winsock2.h> | #include <winsock2.h> | ||||
#include <winsock.h> | #include <winsock.h> | ||||
#include "Windows.h" | #include "Windows.h" | ||||
@@ -103,16 +104,19 @@ | |||||
#include <stdarg.h> | #include <stdarg.h> | ||||
#include "shlwapi.h" | #include "shlwapi.h" | ||||
#include <direct.h> | #include <direct.h> | ||||
#include "sub_inc/mmpa_typedef_win.h" | |||||
#include "sub_inc/mmpa_win.h" | |||||
#include <VersionHelpers.h> | #include <VersionHelpers.h> | ||||
#include <processthreadsapi.h> | #include <processthreadsapi.h> | ||||
#include <Wbemidl.h> | #include <Wbemidl.h> | ||||
#include <iphlpapi.h> | #include <iphlpapi.h> | ||||
#include <synchapi.h> | |||||
#include <sys/types.h> | #include <sys/types.h> | ||||
#include <sys/stat.h> | #include <sys/stat.h> | ||||
#include "securec.h" | |||||
#include "sub_inc/mmpa_typedef_win.h" | |||||
#include "sub_inc/mmpa_win.h" | |||||
#pragma comment(lib, "ws2_32.lib") | #pragma comment(lib, "ws2_32.lib") | ||||
#pragma comment(lib, "mswsock.lib") | #pragma comment(lib, "mswsock.lib") | ||||
#pragma comment(lib, "Kernel32.lib") | #pragma comment(lib, "Kernel32.lib") | ||||
@@ -30,18 +30,26 @@ typedef pthread_t mmThread; | |||||
typedef pthread_mutex_t mmMutex_t; | typedef pthread_mutex_t mmMutex_t; | ||||
typedef pthread_cond_t mmCond; | typedef pthread_cond_t mmCond; | ||||
typedef pthread_mutex_t mmMutexFC; | typedef pthread_mutex_t mmMutexFC; | ||||
typedef pthread_rwlock_t mmRWLock_t; | |||||
typedef signed int mmProcess; | typedef signed int mmProcess; | ||||
typedef int mmPollHandle; | typedef int mmPollHandle; | ||||
typedef int mmPipeHandle; | typedef int mmPipeHandle; | ||||
typedef int mmFileHandle; | |||||
typedef int mmComPletionKey; | typedef int mmComPletionKey; | ||||
typedef int mmCompletionHandle; | typedef int mmCompletionHandle; | ||||
typedef int mmErrorMsg; | |||||
typedef int mmFd_t; | |||||
typedef VOID *mmExitCode; | typedef VOID *mmExitCode; | ||||
typedef key_t mmKey_t; | typedef key_t mmKey_t; | ||||
typedef int mmMsgid; | typedef int mmMsgid; | ||||
typedef struct dirent mmDirent; | typedef struct dirent mmDirent; | ||||
typedef struct shmid_ds mmshmId_ds; | |||||
typedef int (*mmFilter)(const mmDirent *entry); | typedef int (*mmFilter)(const mmDirent *entry); | ||||
typedef int (*mmSort)(const mmDirent **a, const mmDirent **b); | typedef int (*mmSort)(const mmDirent **a, const mmDirent **b); | ||||
typedef size_t mmSize_t; | |||||
typedef off_t mmOfft_t; | |||||
typedef pid_t mmPid_t; | |||||
typedef VOID *(*userProcFunc)(VOID *pulArg); | typedef VOID *(*userProcFunc)(VOID *pulArg); | ||||
@@ -50,6 +58,16 @@ typedef struct { | |||||
VOID *pulArg; // Callback function parameters | VOID *pulArg; // Callback function parameters | ||||
} mmUserBlock_t; | } mmUserBlock_t; | ||||
typedef struct { | |||||
const char *dli_fname; | |||||
void *dli_fbase; | |||||
const char *dli_sname; | |||||
void *dli_saddr; | |||||
size_t dli_size; /* ELF only */ | |||||
int dli_bind; /* ELF only */ | |||||
int dli_type; | |||||
} mmDlInfo; | |||||
typedef struct { | typedef struct { | ||||
int wSecond; // Seconds. [0-60] (1 leap second) | int wSecond; // Seconds. [0-60] (1 leap second) | ||||
int wMinute; // Minutes. [0-59] | int wMinute; // Minutes. [0-59] | ||||
@@ -73,6 +91,7 @@ typedef pthread_key_t mmThreadKey; | |||||
typedef int mmOverLap; | typedef int mmOverLap; | ||||
typedef ssize_t mmSsize_t; | typedef ssize_t mmSsize_t; | ||||
typedef size_t mmSize; // size | |||||
typedef struct { | typedef struct { | ||||
UINT32 createFlag; | UINT32 createFlag; | ||||
@@ -201,6 +220,17 @@ typedef struct { | |||||
#define M_RDWR O_RDWR | #define M_RDWR O_RDWR | ||||
#define M_CREAT O_CREAT | #define M_CREAT O_CREAT | ||||
#define M_BINARY O_RDONLY | #define M_BINARY O_RDONLY | ||||
#define M_TRUNC O_TRUNC | |||||
#define M_IRWXU S_IRWXU | |||||
#define M_IN_CREATE IN_CREATE | |||||
#define M_IN_CLOSE_WRITE IN_CLOSE_WRITE | |||||
#define M_IN_IGNORED IN_IGNORED | |||||
#define M_OUT_CREATE IN_CREATE | |||||
#define M_OUT_CLOSE_WRITE IN_CLOSE_WRITE | |||||
#define M_OUT_IGNORED IN_IGNORED | |||||
#define M_OUT_ISDIR IN_ISDIR | |||||
#define M_IREAD S_IREAD | #define M_IREAD S_IREAD | ||||
#define M_IRUSR S_IRUSR | #define M_IRUSR S_IRUSR | ||||
@@ -236,13 +266,20 @@ typedef struct { | |||||
#define MMPA_OPTIONAL_ARGUMENT 2 | #define MMPA_OPTIONAL_ARGUMENT 2 | ||||
#define MMPA_MAX_PATH PATH_MAX | #define MMPA_MAX_PATH PATH_MAX | ||||
#define M_NAME_MAX MAX_FNAME | |||||
#define M_F_OK F_OK | #define M_F_OK F_OK | ||||
#define M_R_OK R_OK | #define M_R_OK R_OK | ||||
#define M_W_OK W_OK | #define M_W_OK W_OK | ||||
#define MMPA_STDIN STDIN_FILENO | |||||
#define MMPA_STDOUT STDOUT_FILENO | |||||
#define MMPA_STDERR STDERR_FILENO | |||||
#define MMPA_RTLD_NOW RTLD_NOW | #define MMPA_RTLD_NOW RTLD_NOW | ||||
#define MMPA_RTLD_GLOBAL RTLD_GLOBAL | #define MMPA_RTLD_GLOBAL RTLD_GLOBAL | ||||
#define MMPA_RTLD_LAZY RTLD_LAZY | |||||
#define MMPA_RTLD_NODELETE RTLD_NODELETE | |||||
#define MMPA_DL_EXT_NAME ".so" | #define MMPA_DL_EXT_NAME ".so" | ||||
@@ -250,6 +287,7 @@ extern INT32 mmCreateTask(mmThread *threadHandle, mmUserBlock_t *funcBlock); | |||||
extern INT32 mmJoinTask(mmThread *threadHandle); | extern INT32 mmJoinTask(mmThread *threadHandle); | ||||
extern INT32 mmMutexInit(mmMutex_t *mutex); | extern INT32 mmMutexInit(mmMutex_t *mutex); | ||||
extern INT32 mmMutexLock(mmMutex_t *mutex); | extern INT32 mmMutexLock(mmMutex_t *mutex); | ||||
extern INT32 mmMutexTryLock(mmMutex_t *mutex); | |||||
extern INT32 mmMutexUnLock(mmMutex_t *mutex); | extern INT32 mmMutexUnLock(mmMutex_t *mutex); | ||||
extern INT32 mmMutexDestroy(mmMutex_t *mutex); | extern INT32 mmMutexDestroy(mmMutex_t *mutex); | ||||
extern INT32 mmCondInit(mmCond *cond); | extern INT32 mmCondInit(mmCond *cond); | ||||
@@ -257,6 +295,14 @@ extern INT32 mmCondLockInit(mmMutexFC *mutex); | |||||
extern INT32 mmCondLock(mmMutexFC *mutex); | extern INT32 mmCondLock(mmMutexFC *mutex); | ||||
extern INT32 mmCondUnLock(mmMutexFC *mutex); | extern INT32 mmCondUnLock(mmMutexFC *mutex); | ||||
extern INT32 mmCondLockDestroy(mmMutexFC *mutex); | extern INT32 mmCondLockDestroy(mmMutexFC *mutex); | ||||
extern INT32 mmRWLockInit(mmRWLock_t *rwLock); | |||||
extern INT32 mmRWLockRDLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmRWLockTryRDLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmRWLockWRLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmRWLockTryWRLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmRDLockUnLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmWRLockUnLock(mmRWLock_t *rwLock); | |||||
extern INT32 mmRWLockDestroy(mmRWLock_t *rwLock); | |||||
extern INT32 mmCondWait(mmCond *cond, mmMutexFC *mutex); | extern INT32 mmCondWait(mmCond *cond, mmMutexFC *mutex); | ||||
extern INT32 mmCondTimedWait(mmCond *cond, mmMutexFC *mutex, UINT32 milliSecond); | extern INT32 mmCondTimedWait(mmCond *cond, mmMutexFC *mutex, UINT32 milliSecond); | ||||
extern INT32 mmCondNotify(mmCond *cond); | extern INT32 mmCondNotify(mmCond *cond); | ||||
@@ -266,6 +312,7 @@ extern INT32 mmGetPid(); | |||||
extern INT32 mmGetTid(); | extern INT32 mmGetTid(); | ||||
extern INT32 mmGetPidHandle(mmProcess *processHandle); | extern INT32 mmGetPidHandle(mmProcess *processHandle); | ||||
extern INT32 mmGetLocalTime(mmSystemTime_t *sysTime); | extern INT32 mmGetLocalTime(mmSystemTime_t *sysTime); | ||||
extern INT32 mmGetSystemTime(mmSystemTime_t *sysTime); | |||||
extern INT32 mmSemInit(mmSem_t *sem, UINT32 value); | extern INT32 mmSemInit(mmSem_t *sem, UINT32 value); | ||||
extern INT32 mmSemWait(mmSem_t *sem); | extern INT32 mmSemWait(mmSem_t *sem); | ||||
@@ -273,7 +320,9 @@ extern INT32 mmSemPost(mmSem_t *sem); | |||||
extern INT32 mmSemDestroy(mmSem_t *sem); | extern INT32 mmSemDestroy(mmSem_t *sem); | ||||
extern INT32 mmOpen(const CHAR *pathName, INT32 flags); | extern INT32 mmOpen(const CHAR *pathName, INT32 flags); | ||||
extern INT32 mmOpen2(const CHAR *pathName, INT32 flags, MODE mode); | extern INT32 mmOpen2(const CHAR *pathName, INT32 flags, MODE mode); | ||||
extern FILE *mmPopen(CHAR *command, CHAR *type); | |||||
extern INT32 mmClose(INT32 fd); | extern INT32 mmClose(INT32 fd); | ||||
extern INT32 mmPclose(FILE *stream); | |||||
extern mmSsize_t mmWrite(INT32 fd, VOID *buf, UINT32 bufLen); | extern mmSsize_t mmWrite(INT32 fd, VOID *buf, UINT32 bufLen); | ||||
extern mmSsize_t mmRead(INT32 fd, VOID *buf, UINT32 bufLen); | extern mmSsize_t mmRead(INT32 fd, VOID *buf, UINT32 bufLen); | ||||
extern mmSockHandle mmSocket(INT32 sockFamily, INT32 type, INT32 protocol); | extern mmSockHandle mmSocket(INT32 sockFamily, INT32 type, INT32 protocol); | ||||
@@ -284,9 +333,22 @@ extern INT32 mmConnect(mmSockHandle sockFd, mmSockAddr *addr, mmSocklen_t addrLe | |||||
extern INT32 mmCloseSocket(mmSockHandle sockFd); | extern INT32 mmCloseSocket(mmSockHandle sockFd); | ||||
extern mmSsize_t mmSocketSend(mmSockHandle sockFd, VOID *sendBuf, INT32 sendLen, INT32 sendFlag); | extern mmSsize_t mmSocketSend(mmSockHandle sockFd, VOID *sendBuf, INT32 sendLen, INT32 sendFlag); | ||||
extern mmSsize_t mmSocketRecv(mmSockHandle sockFd, VOID *recvBuf, INT32 recvLen, INT32 recvFlag); | extern mmSsize_t mmSocketRecv(mmSockHandle sockFd, VOID *recvBuf, INT32 recvLen, INT32 recvFlag); | ||||
extern INT32 mmSocketSendTo(mmSockHandle sockFd, | |||||
VOID *sendMsg, | |||||
INT32 sendLen, | |||||
UINT32 sendFlag, | |||||
const mmSockAddr* addr, | |||||
INT32 tolen); | |||||
extern mmSsize_t mmSocketRecvFrom(mmSockHandle sockFd, | |||||
VOID *recvBuf, | |||||
mmSize recvLen, | |||||
UINT32 recvFlag, | |||||
mmSockAddr* addr, | |||||
mmSocklen_t *FromLen); | |||||
extern INT32 mmSAStartup(); | extern INT32 mmSAStartup(); | ||||
extern INT32 mmSACleanup(); | extern INT32 mmSACleanup(); | ||||
extern VOID *mmDlopen(const CHAR *fileName, INT32 mode); | extern VOID *mmDlopen(const CHAR *fileName, INT32 mode); | ||||
extern INT32 mmDladdr(VOID *addr, mmDlInfo *info); | |||||
extern VOID *mmDlsym(VOID *handle, CHAR *funcName); | extern VOID *mmDlsym(VOID *handle, CHAR *funcName); | ||||
extern INT32 mmDlclose(VOID *handle); | extern INT32 mmDlclose(VOID *handle); | ||||
extern CHAR *mmDlerror(); | extern CHAR *mmDlerror(); | ||||
@@ -294,6 +356,7 @@ extern INT32 mmCreateAndSetTimer(mmTimer *timerHandle, mmUserBlock_t *timerBlock | |||||
extern INT32 mmDeleteTimer(mmTimer timerHandle); | extern INT32 mmDeleteTimer(mmTimer timerHandle); | ||||
extern INT32 mmStatGet(const CHAR *path, mmStat_t *buffer); | extern INT32 mmStatGet(const CHAR *path, mmStat_t *buffer); | ||||
extern INT32 mmStat64Get(const CHAR *path, mmStat64_t *buffer); | extern INT32 mmStat64Get(const CHAR *path, mmStat64_t *buffer); | ||||
extern INT32 mmFStatGet(INT32 fd, mmStat_t *buffer); | |||||
extern INT32 mmMkdir(const CHAR *pathName, mmMode_t mode); | extern INT32 mmMkdir(const CHAR *pathName, mmMode_t mode); | ||||
extern INT32 mmSleep(UINT32 milliSecond); | extern INT32 mmSleep(UINT32 milliSecond); | ||||
@@ -337,6 +400,7 @@ extern VOID mmCloseCompletionPort(mmCompletionHandle handle); | |||||
extern INT32 mmPoll(mmPollfd *fds, INT32 fdCount, INT32 timeout, mmCompletionHandle handleIOCP, pmmPollData polledData, | extern INT32 mmPoll(mmPollfd *fds, INT32 fdCount, INT32 timeout, mmCompletionHandle handleIOCP, pmmPollData polledData, | ||||
mmPollBack pollBack); | mmPollBack pollBack); | ||||
extern INT32 mmGetErrorCode(); | extern INT32 mmGetErrorCode(); | ||||
extern CHAR *mmGetErrorFormatMessage(mmErrorMsg errnum, CHAR *buf, mmSize size); | |||||
extern INT32 mmGetTimeOfDay(mmTimeval *timeVal, mmTimezone *timeZone); | extern INT32 mmGetTimeOfDay(mmTimeval *timeVal, mmTimezone *timeZone); | ||||
extern mmTimespec mmGetTickCount(); | extern mmTimespec mmGetTickCount(); | ||||
extern INT32 mmGetRealPath(CHAR *path, CHAR *realPath); | extern INT32 mmGetRealPath(CHAR *path, CHAR *realPath); | ||||
@@ -382,6 +446,7 @@ extern INT32 mmTlsDelete(mmThreadKey key); | |||||
extern INT32 mmGetOsType(); | extern INT32 mmGetOsType(); | ||||
extern INT32 mmFsync(mmProcess fd); | extern INT32 mmFsync(mmProcess fd); | ||||
extern INT32 mmFsync2(INT32 fd); | |||||
extern INT32 mmChdir(const CHAR *path); | extern INT32 mmChdir(const CHAR *path); | ||||
extern INT32 mmUmask(INT32 pmode); | extern INT32 mmUmask(INT32 pmode); | ||||
extern INT32 mmThreadKill(mmThread id); | extern INT32 mmThreadKill(mmThread id); | ||||
@@ -439,6 +504,10 @@ extern INT32 mmCreateProcess(const CHAR *fileName, const mmArgvEnv *env, const c | |||||
extern INT32 mmCreateTaskWithThreadAttr(mmThread *threadHandle, const mmUserBlock_t *funcBlock, | extern INT32 mmCreateTaskWithThreadAttr(mmThread *threadHandle, const mmUserBlock_t *funcBlock, | ||||
const mmThreadAttr *threadAttr); | const mmThreadAttr *threadAttr); | ||||
extern mmFileHandle mmShmOpen(const CHAR *name, INT32 oflag, mmMode_t mode); | |||||
extern INT32 mmShmUnlink(const CHAR *name); | |||||
extern VOID *mmMmap(mmFd_t fd, mmSize_t size, mmOfft_t offset, mmFd_t *extra, INT32 prot, INT32 flags); | |||||
extern INT32 mmMunMap(VOID *data, mmSize_t size, mmFd_t *extra); | |||||
#define MMPA_DLL_API | #define MMPA_DLL_API | ||||
#ifdef __cplusplus | #ifdef __cplusplus | ||||
@@ -79,6 +79,8 @@ typedef long LONG; | |||||
#define MMPA_THREAD_SCHED_OTHER SCHED_OTHER | #define MMPA_THREAD_SCHED_OTHER SCHED_OTHER | ||||
#define MMPA_THREAD_MIN_STACK_SIZE PTHREAD_STACK_MIN | #define MMPA_THREAD_MIN_STACK_SIZE PTHREAD_STACK_MIN | ||||
#define MM_MUTEX_INITIALIZER PTHREAD_MUTEX_INITIALIZER | |||||
#define MMPA_MAX_NI 19 | #define MMPA_MAX_NI 19 | ||||
#define MMPA_MIN_NI (-20) | #define MMPA_MIN_NI (-20) | ||||
@@ -86,6 +88,7 @@ typedef long LONG; | |||||
#define EN_ERR 1 | #define EN_ERR 1 | ||||
#define EN_ERROR (-1) | #define EN_ERROR (-1) | ||||
#define EN_INVALID_PARAM (-2) | #define EN_INVALID_PARAM (-2) | ||||
#define EN_TIMEOUT (-3) | |||||
#ifdef __cplusplus | #ifdef __cplusplus | ||||
#if __cplusplus | #if __cplusplus | ||||
@@ -1,4 +1,4 @@ | |||||
/** | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | * Copyright 2019-2020 Huawei Technologies Co., Ltd | ||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
@@ -35,6 +35,7 @@ extern "C" { | |||||
#define EN_ERR 1 | #define EN_ERR 1 | ||||
#define EN_ERROR (-1) | #define EN_ERROR (-1) | ||||
#define EN_INVALID_PARAM (-2) | #define EN_INVALID_PARAM (-2) | ||||
#define EN_TIMEOUT (-3) | |||||
#define HANDLE_INVALID_VALUE (-1) | #define HANDLE_INVALID_VALUE (-1) | ||||
#define INVALID_SOCKET_HANDLE INVALID_SOCKET | #define INVALID_SOCKET_HANDLE INVALID_SOCKET | ||||
@@ -60,6 +61,7 @@ extern "C" { | |||||
#define MMPA_MIDDLE_NI 5 | #define MMPA_MIDDLE_NI 5 | ||||
#define MMPA_LOW_NI (-5) | #define MMPA_LOW_NI (-5) | ||||
#define MMPA_MIN_NI (-20) | #define MMPA_MIN_NI (-20) | ||||
#define MMPA_MAX_FILE 128 | |||||
#define MMPA_MAX_THREAD_PIO 99 | #define MMPA_MAX_THREAD_PIO 99 | ||||
#define MMPA_MIDDLE_THREAD_PIO 66 | #define MMPA_MIDDLE_THREAD_PIO 66 | ||||
@@ -71,6 +73,8 @@ extern "C" { | |||||
#define MMPA_THREAD_SCHED_OTHER 0 | #define MMPA_THREAD_SCHED_OTHER 0 | ||||
#define MMPA_THREAD_MIN_STACK_SIZE 0 | #define MMPA_THREAD_MIN_STACK_SIZE 0 | ||||
#define MM_MUTEX_INITIALIZER NULL | |||||
#ifdef __cplusplus | #ifdef __cplusplus | ||||
#if __cplusplus | #if __cplusplus | ||||
} | } | ||||
@@ -1,4 +1,4 @@ | |||||
/** | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | * Copyright 2019-2020 Huawei Technologies Co., Ltd | ||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
@@ -43,8 +43,9 @@ typedef HANDLE mmThread; | |||||
typedef HANDLE mmProcess; | typedef HANDLE mmProcess; | ||||
typedef HANDLE mmPollHandle; | typedef HANDLE mmPollHandle; | ||||
typedef HANDLE mmPipeHandle; | typedef HANDLE mmPipeHandle; | ||||
typedef HANDLE mmFileHandle; | |||||
typedef HANDLE mmCompletionHandle; | typedef HANDLE mmCompletionHandle; | ||||
typedef HANDLE mmFd_t; | |||||
typedef CRITICAL_SECTION mmMutexFC; | typedef CRITICAL_SECTION mmMutexFC; | ||||
typedef CONDITION_VARIABLE mmCond; | typedef CONDITION_VARIABLE mmCond; | ||||
@@ -59,15 +60,22 @@ typedef SYSTEMTIME mmSystemTime_t; | |||||
typedef HANDLE mmSem_t; | typedef HANDLE mmSem_t; | ||||
typedef SOCKET mmSockHandle; | typedef SOCKET mmSockHandle; | ||||
typedef SRWLOCK mmRWLock_t; | |||||
typedef struct sockaddr mmSockAddr; | typedef struct sockaddr mmSockAddr; | ||||
typedef int mmSocklen_t; | typedef int mmSocklen_t; | ||||
typedef int mmSemTimeout_t; | typedef int mmSemTimeout_t; | ||||
typedef long mmAtomicType; | typedef long mmAtomicType; | ||||
typedef DWORD mmExitCode; | typedef DWORD mmExitCode; | ||||
typedef DWORD mmErrorMsg; | |||||
typedef int mmKey_t; | typedef int mmKey_t; | ||||
typedef HANDLE mmMsgid; | typedef HANDLE mmMsgid; | ||||
typedef long int mmOfft_t; | |||||
typedef int mmPid_t; | |||||
typedef INT32 mmSsize_t; | typedef INT32 mmSsize_t; | ||||
typedef int mmSize; // size | |||||
typedef size_t mmSize_t; | |||||
typedef VOID mmshmId_ds; | |||||
typedef enum { | typedef enum { | ||||
DT_DIR = FILE_ATTRIBUTE_DIRECTORY, | DT_DIR = FILE_ATTRIBUTE_DIRECTORY, | ||||
@@ -181,6 +189,16 @@ typedef struct { | |||||
ULONGLONG availSize; | ULONGLONG availSize; | ||||
} mmDiskSize; | } mmDiskSize; | ||||
typedef struct { | |||||
const char *dli_fname; | |||||
void *dli_fbase; | |||||
const char *dli_sname; | |||||
void *dli_saddr; | |||||
size_t dli_size; /* ELF only */ | |||||
int dli_bind; /* ELF only */ | |||||
int dli_type; | |||||
} mmDlInfo; | |||||
typedef struct { | typedef struct { | ||||
char addr[MMPA_MACINFO_DEFAULT_SIZE]; // ex:aa-bb-cc-dd-ee-ff\0 | char addr[MMPA_MACINFO_DEFAULT_SIZE]; // ex:aa-bb-cc-dd-ee-ff\0 | ||||
} mmMacInfo; | } mmMacInfo; | ||||
@@ -223,8 +241,10 @@ typedef VOID (*mmPf)(VOID); | |||||
#define M_RDONLY _O_RDONLY | #define M_RDONLY _O_RDONLY | ||||
#define M_WRONLY _O_WRONLY | #define M_WRONLY _O_WRONLY | ||||
#define M_RDWR _O_RDWR | #define M_RDWR _O_RDWR | ||||
#define M_IRWXU _O_RDWR | |||||
#define M_CREAT _O_CREAT | #define M_CREAT _O_CREAT | ||||
#define M_BINARY _O_BINARY | #define M_BINARY _O_BINARY | ||||
#define M_TRUNC _O_TRUNC | |||||
#define M_IREAD _S_IREAD | #define M_IREAD _S_IREAD | ||||
#define M_IRUSR _S_IREAD | #define M_IRUSR _S_IREAD | ||||
@@ -232,6 +252,15 @@ typedef VOID (*mmPf)(VOID); | |||||
#define M_IWUSR _S_IWRITE | #define M_IWUSR _S_IWRITE | ||||
#define M_IXUSR 0 | #define M_IXUSR 0 | ||||
#define M_IN_CREATE FILE_NOTIFY_CHANGE_FILE_NAME | FILE_NOTIFY_CHANGE_DIR_NAME | |||||
#define M_IN_CLOSE_WRITE FILE_NOTIFY_CHANGE_LAST_WRITE | |||||
#define M_IN_IGNORED FILE_NOTIFY_CHANGE_FILE_NAME | FILE_NOTIFY_CHANGE_DIR_NAME | |||||
#define M_OUT_CREATE 0x00000100 | |||||
#define M_OUT_CLOSE_WRITE 0x00000008 | |||||
#define M_OUT_IGNORED 0x00008000 | |||||
#define M_OUT_ISDIR 0x40000000 | |||||
#define M_MSG_CREAT 1 | #define M_MSG_CREAT 1 | ||||
#define M_MSG_EXCL 2 | #define M_MSG_EXCL 2 | ||||
#define M_MSG_NOWAIT 3 | #define M_MSG_NOWAIT 3 | ||||
@@ -251,6 +280,16 @@ typedef VOID (*mmPf)(VOID); | |||||
#define M_UMASK_GRPEXEC 0 | #define M_UMASK_GRPEXEC 0 | ||||
#define M_UMASK_OTHEXEC 0 | #define M_UMASK_OTHEXEC 0 | ||||
#define DT_UNKNOWN 0 | |||||
#define DT_FIFO 1 | |||||
#define DT_CHR 2 | |||||
#define DT_DIR 4 | |||||
#define DT_BLK 6 | |||||
#define DT_REG 8 | |||||
#define DT_LNK 10 | |||||
#define DT_SOCK 12 | |||||
#define DT_WHT 14 | |||||
#define mmConstructor(x) __declspec(allocate(".CRT$XCU")) mmPf con = x | #define mmConstructor(x) __declspec(allocate(".CRT$XCU")) mmPf con = x | ||||
#define mmDestructor(x) __declspec(allocate(".CRT$XPU")) mmPf de = x | #define mmDestructor(x) __declspec(allocate(".CRT$XPU")) mmPf de = x | ||||
@@ -269,13 +308,20 @@ typedef VOID (*mmPf)(VOID); | |||||
#define MMPA_EMSG "" | #define MMPA_EMSG "" | ||||
#define MMPA_MAX_PATH MAX_PATH | #define MMPA_MAX_PATH MAX_PATH | ||||
#define M_NAME_MAX _MAX_FNAME | |||||
#define M_F_OK 0 | #define M_F_OK 0 | ||||
#define M_W_OK 2 | #define M_W_OK 2 | ||||
#define M_R_OK 4 | #define M_R_OK 4 | ||||
#define MMPA_STDIN stdin | |||||
#define MMPA_STDOUT stdout | |||||
#define MMPA_STDERR stderr | |||||
#define MMPA_RTLD_NOW 0 | #define MMPA_RTLD_NOW 0 | ||||
#define MMPA_RTLD_GLOBAL 0 | #define MMPA_RTLD_GLOBAL 0 | ||||
#define MMPA_RTLD_LAZY 0 | |||||
#define MMPA_RTLD_NODELETE 0 | |||||
#define MMPA_DL_EXT_NAME ".dll" | #define MMPA_DL_EXT_NAME ".dll" | ||||
@@ -285,6 +331,7 @@ _declspec(dllexport) INT32 mmCreateTask(mmThread *threadHandle, mmUserBlock_t *f | |||||
_declspec(dllexport) INT32 mmJoinTask(mmThread *threadHandle); | _declspec(dllexport) INT32 mmJoinTask(mmThread *threadHandle); | ||||
_declspec(dllexport) INT32 mmMutexInit(mmMutex_t *mutex); | _declspec(dllexport) INT32 mmMutexInit(mmMutex_t *mutex); | ||||
_declspec(dllexport) INT32 mmMutexLock(mmMutex_t *mutex); | _declspec(dllexport) INT32 mmMutexLock(mmMutex_t *mutex); | ||||
_declspec(dllexport) INT32 mmMutexTryLock(mmMutex_t *mutex); | |||||
_declspec(dllexport) INT32 mmMutexUnLock(mmMutex_t *mutex); | _declspec(dllexport) INT32 mmMutexUnLock(mmMutex_t *mutex); | ||||
_declspec(dllexport) INT32 mmMutexDestroy(mmMutex_t *mutex); | _declspec(dllexport) INT32 mmMutexDestroy(mmMutex_t *mutex); | ||||
_declspec(dllexport) INT32 mmCondInit(mmCond *cond); | _declspec(dllexport) INT32 mmCondInit(mmCond *cond); | ||||
@@ -292,6 +339,14 @@ _declspec(dllexport) INT32 mmCondLockInit(mmMutexFC *mutex); | |||||
_declspec(dllexport) INT32 mmCondLock(mmMutexFC *mutex); | _declspec(dllexport) INT32 mmCondLock(mmMutexFC *mutex); | ||||
_declspec(dllexport) INT32 mmCondUnLock(mmMutexFC *mutex); | _declspec(dllexport) INT32 mmCondUnLock(mmMutexFC *mutex); | ||||
_declspec(dllexport) INT32 mmCondLockDestroy(mmMutexFC *mutex); | _declspec(dllexport) INT32 mmCondLockDestroy(mmMutexFC *mutex); | ||||
_declspec(dllexport) INT32 mmRWLockInit(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRWLockRDLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRWLockTryRDLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRWLockWRLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRWLockTryWRLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRDLockUnLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmWRLockUnLock(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmRWLockDestroy(mmRWLock_t *rwLock); | |||||
_declspec(dllexport) INT32 mmCondWait(mmCond *cond, mmMutexFC *mutex); | _declspec(dllexport) INT32 mmCondWait(mmCond *cond, mmMutexFC *mutex); | ||||
_declspec(dllexport) INT32 mmCondTimedWait(mmCond *cond, mmMutexFC *mutex, UINT32 milliSecond); | _declspec(dllexport) INT32 mmCondTimedWait(mmCond *cond, mmMutexFC *mutex, UINT32 milliSecond); | ||||
@@ -302,13 +357,16 @@ _declspec(dllexport) INT32 mmGetPid(VOID); | |||||
_declspec(dllexport) INT32 mmGetTid(VOID); | _declspec(dllexport) INT32 mmGetTid(VOID); | ||||
_declspec(dllexport) INT32 mmGetPidHandle(mmProcess *processHandle); | _declspec(dllexport) INT32 mmGetPidHandle(mmProcess *processHandle); | ||||
_declspec(dllexport) INT32 mmGetLocalTime(mmSystemTime_t *sysTime); | _declspec(dllexport) INT32 mmGetLocalTime(mmSystemTime_t *sysTime); | ||||
_declspec(dllexport) INT32 mmGetSystemTime(mmSystemTime_t *sysTime); | |||||
_declspec(dllexport) INT32 mmSemInit(mmSem_t *sem, UINT32 value); | _declspec(dllexport) INT32 mmSemInit(mmSem_t *sem, UINT32 value); | ||||
_declspec(dllexport) INT32 mmSemWait(mmSem_t *sem); | _declspec(dllexport) INT32 mmSemWait(mmSem_t *sem); | ||||
_declspec(dllexport) INT32 mmSemPost(mmSem_t *sem); | _declspec(dllexport) INT32 mmSemPost(mmSem_t *sem); | ||||
_declspec(dllexport) INT32 mmSemDestroy(mmSem_t *sem); | _declspec(dllexport) INT32 mmSemDestroy(mmSem_t *sem); | ||||
_declspec(dllexport) INT32 mmOpen(const CHAR *pathName, INT32 flags); | _declspec(dllexport) INT32 mmOpen(const CHAR *pathName, INT32 flags); | ||||
_declspec(dllexport) INT32 mmOpen2(const CHAR *pathName, INT32 flags, MODE mode); | _declspec(dllexport) INT32 mmOpen2(const CHAR *pathName, INT32 flags, MODE mode); | ||||
_declspec(dllexport) FILE *mmPopen(CHAR *command, CHAR *type); | |||||
_declspec(dllexport) INT32 mmClose(INT32 fd); | _declspec(dllexport) INT32 mmClose(INT32 fd); | ||||
_declspec(dllexport) INT32 mmPclose(FILE *stream); | |||||
_declspec(dllexport) mmSsize_t mmWrite(INT32 fd, VOID *buf, UINT32 bufLen); | _declspec(dllexport) mmSsize_t mmWrite(INT32 fd, VOID *buf, UINT32 bufLen); | ||||
_declspec(dllexport) mmSsize_t mmRead(INT32 fd, VOID *buf, UINT32 bufLen); | _declspec(dllexport) mmSsize_t mmRead(INT32 fd, VOID *buf, UINT32 bufLen); | ||||
_declspec(dllexport) mmSockHandle mmSocket(INT32 sockFamily, INT32 type, INT32 protocol); | _declspec(dllexport) mmSockHandle mmSocket(INT32 sockFamily, INT32 type, INT32 protocol); | ||||
@@ -319,9 +377,22 @@ _declspec(dllexport) INT32 mmConnect(mmSockHandle sockFd, mmSockAddr *addr, mmSo | |||||
_declspec(dllexport) INT32 mmCloseSocket(mmSockHandle sockFd); | _declspec(dllexport) INT32 mmCloseSocket(mmSockHandle sockFd); | ||||
_declspec(dllexport) mmSsize_t mmSocketRecv(mmSockHandle sockFd, VOID *recvBuf, INT32 recvLen, INT32 recvFlag); | _declspec(dllexport) mmSsize_t mmSocketRecv(mmSockHandle sockFd, VOID *recvBuf, INT32 recvLen, INT32 recvFlag); | ||||
_declspec(dllexport) mmSsize_t mmSocketSend(mmSockHandle sockFd, VOID *sendBuf, INT32 sendLen, INT32 sendFlag); | _declspec(dllexport) mmSsize_t mmSocketSend(mmSockHandle sockFd, VOID *sendBuf, INT32 sendLen, INT32 sendFlag); | ||||
_declspec(dllexport) INT32 mmSocketSendTo(mmSockHandle sockFd, | |||||
VOID *sendMsg, | |||||
INT32 sendLen, | |||||
UINT32 sendFlag, | |||||
const mmSockAddr* addr, | |||||
INT32 tolen); | |||||
_declspec(dllexport) mmSsize_t mmSocketRecvFrom(mmSockHandle sockFd, | |||||
VOID *recvBuf, | |||||
mmSize recvLen, | |||||
UINT32 recvFlag, | |||||
mmSockAddr* addr, | |||||
mmSocklen_t *FromLen); | |||||
_declspec(dllexport) INT32 mmSAStartup(VOID); | _declspec(dllexport) INT32 mmSAStartup(VOID); | ||||
_declspec(dllexport) INT32 mmSACleanup(VOID); | _declspec(dllexport) INT32 mmSACleanup(VOID); | ||||
_declspec(dllexport) VOID *mmDlopen(const CHAR *fileName, INT mode); | _declspec(dllexport) VOID *mmDlopen(const CHAR *fileName, INT mode); | ||||
_declspec(dllexport) INT32 mmDladdr(VOID *addr, mmDlInfo *info); | |||||
_declspec(dllexport) VOID *mmDlsym(VOID *handle, CHAR *fileName); | _declspec(dllexport) VOID *mmDlsym(VOID *handle, CHAR *fileName); | ||||
_declspec(dllexport) INT32 mmDlclose(VOID *handle); | _declspec(dllexport) INT32 mmDlclose(VOID *handle); | ||||
_declspec(dllexport) CHAR *mmDlerror(VOID); | _declspec(dllexport) CHAR *mmDlerror(VOID); | ||||
@@ -330,6 +401,7 @@ _declspec(dllexport) INT32 | |||||
_declspec(dllexport) INT32 mmDeleteTimer(mmTimer timerHandle); | _declspec(dllexport) INT32 mmDeleteTimer(mmTimer timerHandle); | ||||
_declspec(dllexport) INT32 mmStatGet(const CHAR *path, mmStat_t *buffer); | _declspec(dllexport) INT32 mmStatGet(const CHAR *path, mmStat_t *buffer); | ||||
_declspec(dllexport) INT32 mmStat64Get(const CHAR *path, mmStat64_t *buffer); | _declspec(dllexport) INT32 mmStat64Get(const CHAR *path, mmStat64_t *buffer); | ||||
_declspec(dllexport) INT32 mmFStatGet(INT32 fd, mmStat_t *buffer); | |||||
_declspec(dllexport) INT32 mmMkdir(const CHAR *pathName, mmMode_t mode); | _declspec(dllexport) INT32 mmMkdir(const CHAR *pathName, mmMode_t mode); | ||||
_declspec(dllexport) INT32 mmSleep(UINT32 milliSecond); | _declspec(dllexport) INT32 mmSleep(UINT32 milliSecond); | ||||
_declspec(dllexport) INT32 mmCreateTaskWithAttr(mmThread *threadHandle, mmUserBlock_t *funcBlock); | _declspec(dllexport) INT32 mmCreateTaskWithAttr(mmThread *threadHandle, mmUserBlock_t *funcBlock); | ||||
@@ -371,6 +443,7 @@ _declspec(dllexport) INT32 mmPoll(mmPollfd *fds, INT32 fdCount, INT32 timeout, m | |||||
pmmPollData polledData, mmPollBack pollBack); | pmmPollData polledData, mmPollBack pollBack); | ||||
_declspec(dllexport) INT32 mmGetErrorCode(); | _declspec(dllexport) INT32 mmGetErrorCode(); | ||||
_declspec(dllexport) CHAR *mmGetErrorFormatMessage(mmErrorMsg errnum, CHAR *buf, mmSize size); | |||||
_declspec(dllexport) INT32 mmGetTimeOfDay(mmTimeval *timeVal, mmTimezone *timeZone); | _declspec(dllexport) INT32 mmGetTimeOfDay(mmTimeval *timeVal, mmTimezone *timeZone); | ||||
_declspec(dllexport) mmTimespec mmGetTickCount(); | _declspec(dllexport) mmTimespec mmGetTickCount(); | ||||
_declspec(dllexport) INT32 mmGetRealPath(CHAR *path, CHAR *realPath); | _declspec(dllexport) INT32 mmGetRealPath(CHAR *path, CHAR *realPath); | ||||
@@ -407,7 +480,7 @@ _declspec(dllexport) INT32 mmTlsDelete(mmThreadKey key); | |||||
_declspec(dllexport) INT32 mmGetOsType(); | _declspec(dllexport) INT32 mmGetOsType(); | ||||
_declspec(dllexport) INT32 mmFsync(mmProcess fd); | _declspec(dllexport) INT32 mmFsync(mmProcess fd); | ||||
_declspec(dllexport) INT32 mmFsync2(INT32 fd); | |||||
_declspec(dllexport) INT32 mmChdir(const CHAR *path); | _declspec(dllexport) INT32 mmChdir(const CHAR *path); | ||||
_declspec(dllexport) INT32 mmUmask(INT32 pmode); | _declspec(dllexport) INT32 mmUmask(INT32 pmode); | ||||
_declspec(dllexport) INT32 mmWaitPid(mmProcess pid, INT32 *status, INT32 options); | _declspec(dllexport) INT32 mmWaitPid(mmProcess pid, INT32 *status, INT32 options); | ||||
@@ -455,7 +528,10 @@ _declspec(dllexport) INT32 | |||||
_declspec(dllexport) INT32 | _declspec(dllexport) INT32 | ||||
mmCreateTaskWithThreadAttr(mmThread *threadHandle, const mmUserBlock_t *funcBlock, const mmThreadAttr *threadAttr); | mmCreateTaskWithThreadAttr(mmThread *threadHandle, const mmUserBlock_t *funcBlock, const mmThreadAttr *threadAttr); | ||||
_declspec(dllexport) mmFileHandle mmShmOpen(const CHAR *name, INT32 oflag, mmMode_t mode); | |||||
_declspec(dllexport) INT32 mmShmUnlink(const CHAR *name); | |||||
_declspec(dllexport) VOID *mmMmap(mmFd_t fd, mmSize_t size, mmOfft_t offset, mmFd_t *extra, INT32 prot, INT32 flags); | |||||
_declspec(dllexport) INT32 mmMunMap(VOID *data, mmSize_t size, mmFd_t *extra); | |||||
#ifdef __cplusplus | #ifdef __cplusplus | ||||
#if __cplusplus | #if __cplusplus | ||||
} | } | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,77 +0,0 @@ | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* | |||||
* Licensed under the Apache License, Version 2.0 (the "License"); | |||||
* you may not use this file except in compliance with the License. | |||||
* You may obtain a copy of the License at | |||||
* | |||||
* http://www.apache.org/licenses/LICENSE-2.0 | |||||
* | |||||
* Unless required by applicable law or agreed to in writing, software | |||||
* distributed under the License is distributed on an "AS IS" BASIS, | |||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||||
* See the License for the specific language governing permissions and | |||||
* limitations under the License. | |||||
*/ | |||||
#ifndef OPS_BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_H_ | |||||
#define BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_H_ | |||||
#include "graph/operator_reg.h" | |||||
namespace ge { | |||||
/** | |||||
* @brief Performs max pooling on the input . \n | |||||
* @par Inputs: | |||||
* One input: | |||||
* x: An NC1HWC0 Tensor. Supported type:float16, float32, double, int8, int16, | |||||
* int32, int64, uint8, uint16, qint8 | |||||
* @par Attributes: | |||||
* @li ksize: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the size of the window for each dimension of the input tensor. | |||||
* No default value. | |||||
* @li strides: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the stride of the sliding window for each dimension of | |||||
* the input tensor. No default value. | |||||
* @li padding_mode: A required string. Defaults to "CALCULATED". | |||||
* @li pads:A required list of int8, int16, int32, or int64 values, | |||||
* a data to caculate when padding_mode is "SAME" and "CALCULATED". | |||||
* @li data_format: An optional string. Defaults to "NHWC" . | |||||
* @li global_pooling bool, Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False | |||||
* @li ceil_mode:global_pooling (bool) – (bool) Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False \n | |||||
* @par Outputs: | |||||
* y: A Tensor. Has the same type and format as input "x" . \n | |||||
* @attention Constraints: | |||||
* @li "ksize" is a list that has length 4: ksize[0] = 1 or ksize[3] = 1, | |||||
* ksize[1] * ksize[2] <= 255. | |||||
* @li "stride is a list that has length 4: strides[0] = 1 or strides[3] = 1, | |||||
* strides[1] <= 63, strides[0] >= 1, strides[2] <= 63, strides[2] >= 1. | |||||
* @li "padding" is "SAME" "VALID" or "CACULATE" . | |||||
* @par Third-party framework compatibility | |||||
* Compatible with the TensorFlow operator MaxPool. | |||||
*/ | |||||
REG_OP(MaxPoolV3) | |||||
.INPUT(x,TensorType({DT_FLOAT16, DT_FLOAT32})) | |||||
.OUTPUT(y, TensorType({DT_FLOAT16, DT_FLOAT32})) | |||||
.REQUIRED_ATTR(ksize, ListInt) | |||||
.REQUIRED_ATTR(strides, ListInt) | |||||
.ATTR(padding_mode, String, "CALCULATED") | |||||
.ATTR(pads, ListInt, {0,0,0,0}) | |||||
.ATTR(data_format, String, "NCHW") | |||||
.ATTR(global_pooling,Bool,false) | |||||
.ATTR(ceil_mode, Bool, false) | |||||
.OP_END_FACTORY_REG(MaxPoolV3) | |||||
} // namespace ge | |||||
#endif // OPS_BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_H_ |
@@ -1,80 +0,0 @@ | |||||
/** | |||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* | |||||
* Licensed under the Apache License, Version 2.0 (the "License"); | |||||
* you may not use this file except in compliance with the License. | |||||
* You may obtain a copy of the License at | |||||
* | |||||
* http://www.apache.org/licenses/LICENSE-2.0 | |||||
* | |||||
* Unless required by applicable law or agreed to in writing, software | |||||
* distributed under the License is distributed on an "AS IS" BASIS, | |||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |||||
* See the License for the specific language governing permissions and | |||||
* limitations under the License. | |||||
*/ | |||||
/*! | |||||
*\file max_pool_v3_grad.h | |||||
*\brief | |||||
*/ | |||||
#ifndef OPS_BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_GRAD_H_ | |||||
#define OPS_BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_GRAD_H_ | |||||
#include "graph/operator_reg.h" | |||||
namespace ge { | |||||
/** | |||||
* @brief Computes gradients of the maxpooling function . \n | |||||
* @par Inputs: | |||||
* @li orig_input: A mutable NC1HWC0 tensor of type RealNumberType. | |||||
* @li orig_output: A mutable NC1HWC0 tensor of type RealNumberTypex. | |||||
* @li grad: A mutable NC1HWC0 tensor of type RealNumberType . \n | |||||
* @par Attributes: | |||||
* @li ksize: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the size of the window for each dimension of the input tensor. | |||||
* No default value. | |||||
* @li strides: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the stride of the sliding window for each dimension of | |||||
* the input tensor. No default value. | |||||
* @li padding_mode: A required string. Defaults to "CALCULATED". | |||||
* @li pads:A required list of int8, int16, int32, or int64 values, | |||||
* a data to caculate when padding_mode is "SAME" and "CALCULATED". | |||||
* @li data_format: An optional string. Defaults to "NHWC" . | |||||
* @li global_pooling bool, Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False | |||||
* @li ceil_mode:global_pooling (bool) – (bool) Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False \n | |||||
* @par Outputs: | |||||
* y: A mutable tensor. Has the same shape and type as "x1" . \n | |||||
* @attention Constraints: | |||||
* @li Computing gradients of global pooling is not supported, which means | |||||
* "ksize < x1". | |||||
* @li "ksize" is in the range [1, 255]. "strides" is in the range [1, 63] | |||||
* @par Third-party framework compatibility | |||||
* Compatible with the TensorFlow operator MaxPoolGrad. | |||||
*/ | |||||
REG_OP(MaxPoolV3Grad) | |||||
.INPUT(orig_input, TensorType::RealNumberType()) | |||||
.INPUT(orig_output, TensorType::RealNumberType()) | |||||
.INPUT(grad, TensorType::RealNumberType()) | |||||
.OUTPUT(out_grad, TensorType::RealNumberType()) | |||||
.REQUIRED_ATTR(ksize, ListInt) | |||||
.REQUIRED_ATTR(strides, ListInt) | |||||
.ATTR(padding_mod, String, "CALCULATED") | |||||
.ATTR(pads, ListInt, {0, 0, 0, 0}) | |||||
.ATTR(data_format, String, "NCHW") | |||||
.ATTR(global_pooling, Bool, false) | |||||
.ATTR(ceil_mode, Bool, false) | |||||
.OP_END_FACTORY_REG(MaxPoolV3Grad) | |||||
} // namespace ge | |||||
#endif // OPS_BUILT_IN_OP_PROTO_INC_MAX_POOL_V3_GRAD_H_ |
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1190,6 +1190,108 @@ REG_OP(MaxPoolGradWithArgmaxV2) | |||||
.ATTR(dilation, ListInt, {1,1,1,1}) | .ATTR(dilation, ListInt, {1,1,1,1}) | ||||
.ATTR(ceil_mode, Bool, false) | .ATTR(ceil_mode, Bool, false) | ||||
.OP_END_FACTORY_REG(MaxPoolGradWithArgmaxV2) | .OP_END_FACTORY_REG(MaxPoolGradWithArgmaxV2) | ||||
} // namespace ge | |||||
#endif // OPS_BUILT_IN_OP_PROTO_INC_NN_POOLING_OPS_H_ | |||||
/** | |||||
* @brief Performs max pooling on the input . \n | |||||
* @par Inputs: | |||||
* One input: | |||||
* x: An NC1HWC0 Tensor. Supported type:float16, float32, double, int8, int16, | |||||
* int32, int64, uint8, uint16, qint8 | |||||
* @par Attributes: | |||||
* @li ksize: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the size of the window for each dimension of the input tensor. | |||||
* No default value. | |||||
* @li strides: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the stride of the sliding window for each dimension of | |||||
* the input tensor. No default value. | |||||
* @li padding_mode: A required string. Defaults to "CALCULATED". | |||||
* @li pads:A required list of int8, int16, int32, or int64 values, | |||||
* a data to caculate when padding_mode is "SAME" and "CALCULATED". | |||||
* @li data_format: An optional string. Defaults to "NHWC" . | |||||
* @li global_pooling bool, Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False | |||||
* @li ceil_mode:global_pooling (bool) – (bool) Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False \n | |||||
* @par Outputs: | |||||
* y: A Tensor. Has the same type and format as input "x" . \n | |||||
* @attention Constraints: | |||||
* @li "ksize" is a list that has length 4: ksize[0] = 1 or ksize[3] = 1, | |||||
* ksize[1] * ksize[2] <= 255. | |||||
* @li "stride is a list that has length 4: strides[0] = 1 or strides[3] = 1, | |||||
* strides[1] <= 63, strides[0] >= 1, strides[2] <= 63, strides[2] >= 1. | |||||
* @li "padding" is "SAME" "VALID" or "CACULATE" . | |||||
* @par Third-party framework compatibility | |||||
* Compatible with the TensorFlow operator MaxPool. | |||||
*/ | |||||
REG_OP(MaxPoolV3) | |||||
.INPUT(x,TensorType({DT_FLOAT16, DT_FLOAT32})) | |||||
.OUTPUT(y, TensorType({DT_FLOAT16, DT_FLOAT32})) | |||||
.REQUIRED_ATTR(ksize, ListInt) | |||||
.REQUIRED_ATTR(strides, ListInt) | |||||
.ATTR(padding_mode, String, "CALCULATED") | |||||
.ATTR(pads, ListInt, {0,0,0,0}) | |||||
.ATTR(data_format, String, "NCHW") | |||||
.ATTR(global_pooling,Bool,false) | |||||
.ATTR(ceil_mode, Bool, false) | |||||
.OP_END_FACTORY_REG(MaxPoolV3) | |||||
/** | |||||
* @brief Computes gradients of the maxpooling function . \n | |||||
* @par Inputs: | |||||
* @li orig_input: A mutable NC1HWC0 tensor of type RealNumberType. | |||||
* @li orig_output: A mutable NC1HWC0 tensor of type RealNumberTypex. | |||||
* @li grad: A mutable NC1HWC0 tensor of type RealNumberType . \n | |||||
* @par Attributes: | |||||
* @li ksize: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the size of the window for each dimension of the input tensor. | |||||
* No default value. | |||||
* @li strides: A required list of int8, int16, int32, or int64 values, | |||||
* specifying the stride of the sliding window for each dimension of | |||||
* the input tensor. No default value. | |||||
* @li padding_mode: A required string. Defaults to "CALCULATED". | |||||
* @li pads:A required list of int8, int16, int32, or int64 values, | |||||
* a data to caculate when padding_mode is "SAME" and "CALCULATED". | |||||
* @li data_format: An optional string. Defaults to "NHWC" . | |||||
* @li global_pooling bool, Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False | |||||
* @li ceil_mode:global_pooling (bool) – (bool) Whether to use the global pooling. | |||||
* If global_pooling = true, kernel size and paddings will be ignored. | |||||
* Default False \n | |||||
* @par Outputs: | |||||
* y: A mutable tensor. Has the same shape and type as "x1" . \n | |||||
* @attention Constraints: | |||||
* @li Computing gradients of global pooling is not supported, which means | |||||
* "ksize < x1". | |||||
* @li "ksize" is in the range [1, 255]. "strides" is in the range [1, 63] | |||||
* @par Third-party framework compatibility | |||||
* Compatible with the TensorFlow operator MaxPoolGrad. | |||||
*/ | |||||
REG_OP(MaxPoolV3Grad) | |||||
.INPUT(orig_input, TensorType::RealNumberType()) | |||||
.INPUT(orig_output, TensorType::RealNumberType()) | |||||
.INPUT(grad, TensorType::RealNumberType()) | |||||
.OUTPUT(out_grad, TensorType::RealNumberType()) | |||||
.REQUIRED_ATTR(ksize, ListInt) | |||||
.REQUIRED_ATTR(strides, ListInt) | |||||
.ATTR(padding_mode, String, "CALCULATED") | |||||
.ATTR(pads, ListInt, {0, 0, 0, 0}) | |||||
.ATTR(data_format, String, "NCHW") | |||||
.ATTR(global_pooling, Bool, false) | |||||
.ATTR(ceil_mode, Bool, false) | |||||
.OP_END_FACTORY_REG(MaxPoolV3Grad) | |||||
} // namespace ge | |||||
#endif // OPS_BUILT_IN_OP_PROTO_INC_NN_POOLING_OPS_H |
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2020 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -343,7 +343,7 @@ REG_OP(AscendPadding) | |||||
*/ | */ | ||||
REG_OP(EmbeddingRankId) | REG_OP(EmbeddingRankId) | ||||
.INPUT(addr_table, TensorType({DT_UINT64})) | .INPUT(addr_table, TensorType({DT_UINT64})) | ||||
.INPUT(index, TensorType({DT_UINT32})) | |||||
.INPUT(index, TensorType({DT_INT64,DT_INT32,DT_UINT64})) | |||||
.OUTPUT(rank_id, TensorType({DT_UINT64})) | .OUTPUT(rank_id, TensorType({DT_UINT64})) | ||||
.ATTR(row_memory, Int, 320) | .ATTR(row_memory, Int, 320) | ||||
.ATTR(mode, String, "mod") | .ATTR(mode, String, "mod") | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||
@@ -1,5 +1,5 @@ | |||||
/** | /** | ||||
* Copyright 2019-2020 Huawei Technologies Co., Ltd | |||||
* Copyright 2019 Huawei Technologies Co., Ltd | |||||
* | * | ||||
* Licensed under the Apache License, Version 2.0 (the "License"); | * Licensed under the Apache License, Version 2.0 (the "License"); | ||||
* you may not use this file except in compliance with the License. | * you may not use this file except in compliance with the License. | ||||