ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ls bin examples models run.sh build gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu onnx_1.7.0_x86_u18_bak scripts dlr_1.10.0_x86_u18 gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu.tar.xz opencv_4.2.0_x86_u18 setup.sh Dockerfile lib output_images test_data dockers libdlr.so README.md tflite_2.8_x86_u18 docs LICENSE requirements_pc_bak.txt tidl_tools entrypoint.sh model-artifacts requirements_pc.txt ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ source ./setup.sh X64 Architecture Installing python packages... Defaulting to user installation because normal site-packages is not writeable Looking in indexes: https://pypi.tuna.tsinghua.edu.cn/simple Collecting git+https://github.com/kumardesappan/caffe2onnx (from -r ./requirements_pc.txt (line 12)) Cloning https://github.com/kumardesappan/caffe2onnx to /tmp/pip-req-build-_xim1adm Running command git clone --filter=blob:none -q https://github.com/kumardesappan/caffe2onnx /tmp/pip-req-build-_xim1adm Resolved https://github.com/kumardesappan/caffe2onnx to commit b7e73feed3bbc5ddbdf25b87af93a2bae596055d Preparing metadata (setup.py) ... done Requirement already satisfied: numpy in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (line 1)) (1.19.5) Requirement already satisfied: pyyaml in /usr/lib/python3/dist-packages (from -r ./requirements_pc.txt (line 2)) (3.12) Requirement already satisfied: protobuf==3.19.4 in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc .txt (line 3)) (3.19.4) Requirement already satisfied: onnx==1.9.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (line 4)) (1.9.0) Requirement already satisfied: tflite in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (line 5)) (2.10.0) Requirement already satisfied: pillow==8.2.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.tx t (line 6)) (8.2.0) Requirement already satisfied: flatbuffers==1.12.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements _pc.txt (line 7)) (1.12) Requirement already satisfied: requests in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (li ne 8)) (2.27.1) Requirement already satisfied: opencv-python==4.2.0.34 in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirem ents_pc.txt (line 9)) (4.2.0.34) Requirement already satisfied: pytest in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (line 10)) (7.0.1) Requirement already satisfied: graphviz in /home/ubuntu/.local/lib/python3.6/site-packages (from -r ./requirements_pc.txt (li ne 11)) (0.8.4) Requirement already satisfied: typing-extensions>=3.6.2.1 in /home/ubuntu/.local/lib/python3.6/site-packages (from onnx==1.9. 0->-r ./requirements_pc.txt (line 4)) (3.7.4.3) Requirement already satisfied: six in /home/ubuntu/.local/lib/python3.6/site-packages (from onnx==1.9.0->-r ./requirements_pc .txt (line 4)) (1.15.0) Requirement already satisfied: charset-normalizer~=2.0.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from requests->- r ./requirements_pc.txt (line 8)) (2.0.12) Requirement already satisfied: idna<4,>=2.5 in /usr/lib/python3/dist-packages (from requests->-r ./requirements_pc.txt (line 8)) (2.6) Requirement already satisfied: certifi>=2017.4.17 in /usr/lib/python3/dist-packages (from requests->-r ./requirements_pc.txt (line 8)) (2018.1.18) Requirement already satisfied: urllib3<1.27,>=1.21.1 in /usr/lib/python3/dist-packages (from requests->-r ./requirements_pc.t xt (line 8)) (1.22) Requirement already satisfied: pluggy<2.0,>=0.12 in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requir ements_pc.txt (line 10)) (1.0.0) Requirement already satisfied: packaging in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requirements_p c.txt (line 10)) (21.3) Requirement already satisfied: iniconfig in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requirements_p c.txt (line 10)) (1.1.1) Requirement already satisfied: importlib-metadata>=0.12 in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r . /requirements_pc.txt (line 10)) (4.8.3) Requirement already satisfied: attrs>=19.2.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requiremen ts_pc.txt (line 10)) (22.2.0) Requirement already satisfied: py>=1.8.2 in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requirements_p c.txt (line 10)) (1.11.0) Requirement already satisfied: tomli>=1.0.0 in /home/ubuntu/.local/lib/python3.6/site-packages (from pytest->-r ./requirement s_pc.txt (line 10)) (1.2.3) Requirement already satisfied: zipp>=0.5 in /home/ubuntu/.local/lib/python3.6/site-packages (from importlib-metadata>=0.12->p ytest->-r ./requirements_pc.txt (line 10)) (3.6.0) Requirement already satisfied: pyparsing!=3.0.5,>=2.0.2 in /home/ubuntu/.local/lib/python3.6/site-packages (from packaging->p ytest->-r ./requirements_pc.txt (line 10)) (3.1.0) Installing python osrt packages... skipping gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu download: found /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-ti dl-tools/gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu Installing:onnxruntime Installing:tflite_2.8 Installing:opencv Installing:dlr ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ls bin docs lib onnx_1.7.0_x86_u18_bak requirements_pc.txt tflite_2.8_x86_u18 build entrypoint.sh libdlr.so opencv_4.2.0_x86_u18 run.sh tidl_tools dlr_1.10.0_x86_u18 examples LICENSE output_images scripts Dockerfile gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu model-artifacts README.md setup.sh dockers gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu.tar.xz models requirements_pc_bak.txt test_data ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ rm -rf build/ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ls bin entrypoint.sh libdlr.so opencv_4.2.0_x86_u18 run.sh tidl_tools dlr_1.10.0_x86_u18 examples LICENSE output_images scripts Dockerfile gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu model-artifacts README.md setup.sh dockers gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu.tar.xz models requirements_pc_bak.txt test_data docs lib onnx_1.7.0_x86_u18_bak requirements_pc.txt tflite_2.8_x86_u18 ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ export SOC=am69a ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ export SOC=am69a ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ export TIDL_TOOLS_PATH=$(pwd)/tidl_tools ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$TIDL_TOOLS_PATH ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ export ARM64_GCC_PATH=$(pwd)/gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ mkdir build && cd build ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/build$ cmake ../examples && make -j && cd .. -- The C compiler identification is GNU 7.5.0 -- The CXX compiler identification is GNU 7.5.0 -- Check for working C compiler: /usr/bin/cc -- Check for working C compiler: /usr/bin/cc -- works -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Detecting C compile features -- Detecting C compile features - done -- Check for working CXX compiler: /usr/bin/c++ -- Check for working CXX compiler: /usr/bin/c++ -- works -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Detecting CXX compile features -- Detecting CXX compile features - done -- Detected processor: x86_64 -- TARGET_DEVICE setting to: am69a -- TARGET_CPU not specicfied using x86 -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = edgeai_tidl_examples -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = edgeai_tidl_examples -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = edgeai_tidl_examples -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = tfl_main -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = tfl_priority_scheduling -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = ort_priority_scheduling -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = edgeai_tidl_examples -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = tidlrt_clasification -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = dlr_main -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- CMAKE_BUILD_TYPE = Release PROJECT_NAME = ort_main -- setting TENSORFLOW_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/tflite_2.8_x86_u18/ -- setting DLR_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/dlr_1.10.0_x86_u18/ -- setting OPENCV_INSTALL_DIR path:/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/tidl_tools/osrt_deps/opencv_4.2.0_x86_u18/ -- Compiling for x86 with am69a config -- Configuring done -- Generating done -- Build files have been written to: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/build Scanning dependencies of target post_process Scanning dependencies of target utils Scanning dependencies of target pre_process Scanning dependencies of target utils_adv [ 3%] Building CXX object osrt_cpp/advanced_examples/utils/CMakeFiles/utils_adv.dir/src/arg_parsing.cpp.o [ 7%] Building CXX object osrt_cpp/advanced_examples/utils/CMakeFiles/utils_adv.dir/src/print_utils.cpp.o [ 11%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/ti_logger.cpp.o [ 14%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/pbPlots.cpp.o [ 18%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/utility_functs.cpp.o [ 22%] Building CXX object osrt_cpp/post_process/CMakeFiles/post_process.dir/post_process.cpp.o [ 25%] Building CXX object osrt_cpp/pre_process/CMakeFiles/pre_process.dir/pre_process.cpp.o [ 29%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/model_info.cpp.o [ 33%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/edgeai_classnames.cpp.o [ 37%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/arg_parsing.cpp.o [ 40%] Building CXX object osrt_cpp/utils/CMakeFiles/utils.dir/src/supportLib.cpp.o [ 44%] Linking CXX static library /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/lib/Release/libutils_adv.a [ 44%] Built target utils_adv [ 48%] Linking CXX static library /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/lib/Release/libpre_process.a [ 48%] Built target pre_process [ 51%] Linking CXX static library /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/lib/Release/libpost_process.a [ 51%] Built target post_process [ 55%] Linking CXX static library /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/lib/Release/libutils.a [ 55%] Built target utils Scanning dependencies of target tfl_priority_scheduling Scanning dependencies of target tfl_main Scanning dependencies of target ort_priority_scheduling Scanning dependencies of target tidlrt_clasification Scanning dependencies of target dlr_main Scanning dependencies of target ort_main [ 59%] Building CXX object tidlrt_cpp/CMakeFiles/tidlrt_clasification.dir/classification.cpp.o [ 62%] Building CXX object osrt_cpp/advanced_examples/ort/CMakeFiles/ort_priority_scheduling.dir/ort_priority_scheduling.cpp.o [ 66%] Building CXX object osrt_cpp/dlr/CMakeFiles/dlr_main.dir/dlr_main.cpp.o [ 70%] Building CXX object osrt_cpp/ort/CMakeFiles/ort_main.dir/onnx_main.cpp.o [ 74%] Building CXX object osrt_cpp/tfl/CMakeFiles/tfl_main.dir/tfl_main.cpp.o [ 77%] Building CXX object osrt_cpp/advanced_examples/tfl/CMakeFiles/tfl_priority_scheduling.dir/tfl_priority_scheduling.cpp.o /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/tidlrt_cpp/classification.cpp: In function ‘void getModelNameromArtifactsDir(char*, char*, char*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/tidlrt_cpp/classification.cpp:155:9: warning: ignoring return value of ‘int fscanf(FILE*, const cha r*, ...)’, declared with attribute warn_unused_result [-Wunused-result] fscanf(fp, "%s", net_name); ~~~~~~^~~~~~~~~~~~~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/tidlrt_cpp/classification.cpp:164:9: warning: ignoring return value of ‘int fscanf(FILE*, const cha r*, ...)’, declared with attribute warn_unused_result [-Wunused-result] fscanf(fp, "%s", io_name); ~~~~~~^~~~~~~~~~~~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/tidlrt_cpp/classification.cpp: In function ‘int32_t TIDLReadBinFromFile(const char*, void*, int32_t )’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/tidlrt_cpp/classification.cpp:174:12: warning: ignoring return value of ‘size_t fread(void*, size_t , size_t, FILE*)’, declared with attribute warn_unused_result [-Wunused-result] fread(addr, size, 1, fptr); ~~~~~^~~~~~~~~~~~~~~~~~~~~ [ 81%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/tidlrt_clasification [ 81%] Built target tidlrt_clasification /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp: In function ‘int tflite::main::getInvok eTime(tflite::main::tfl_model_struct*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:250:58: warning: format ‘%d’ expects arg ument of type ‘int’, but argument 3 has type ‘double’ [-Wformat=] std::sprintf(pre_empt_char, "%d", FLT_MAX); ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:251:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] char *keys[] = {"artifacts_folder", "num_tidl_subgraphs", "debug_level", "priority", "max_pre_empt_delay"}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:251:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:251:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:251:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:251:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:252:146: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] char *values[] = {(char *)arg->modelInfo->m_infConfig.artifactsPath.c_str(), "16", "0", (char *)prior_char, (char *)pre_empt_char}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:252:146: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp: In function ‘void* tflite::main::infer( void*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:439:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] char *keys[] = {"artifacts_folder", "num_tidl_subgraphs", "debug_level", "priority", "max_pre_empt_delay"}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:439:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:439:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:439:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:439:122: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:440:146: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] char *values[] = {(char *)arg->modelInfo->m_infConfig.artifactsPath.c_str(), "16", "0", (char *)prior_char, (char *)pre_empt_char}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/tfl/tfl_priority_scheduling.cpp:440:146: warning: ISO C++ forbids conver ting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp: In function ‘int tflite::main::runInference(tidl::modelInfo::ModelInfo*, tidl::arg_parsing::Settings*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp:255:80: warning: ISO C++ forbids converting a string constant to ‘char*’ [-Wwrite-strings] char *keys[] = {"artifacts_folder", "num_tidl_subgraphs", "debug_level"}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp:255:80: warning: ISO C++ forbids converting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp:255:80: warning: ISO C++ forbids converting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp:256:90: warning: ISO C++ forbids converting a string constant to ‘char*’ [-Wwrite-strings] char *values[] = {(char *)modelInfo->m_infConfig.artifactsPath.c_str(), "16", "0"}; ^ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/tfl/tfl_main.cpp:256:90: warning: ISO C++ forbids converting a string constant to ‘char*’ [-Wwrite-strings] /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp: In function ‘int onnx::main::getInvokeT ime(onnx::main::ort_model_struct*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp:379:50: warning: ignoring return value o f ‘OrtStatus* OrtSessionsOptionsSetDefault_Tidl(c_api_tidl_options*)’, declared with attribute warn_unused_result [-Wunused-result] OrtSessionsOptionsSetDefault_Tidl(options); ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp: In function ‘int onnx::main::runInferen ce(tidl::modelInfo::ModelInfo**, tidl::arg_parsing_adv::Settings*)’: /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp:995:49: warning: ignoring return value o f function declared with attribute warn_unused_result [-Wunused-result] Ort::GetApi().CreateThreadingOptions(&envOpts); ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp:996:53: warning: ignoring return value o f function declared with attribute warn_unused_result [-Wunused-result] Ort::GetApi().SetGlobalInterOpNumThreads(envOpts, 0); ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp:997:47: warning: ignoring return value o f function declared with attribute warn_unused_result [-Wunused-result] Ort::GetApi().SetGlobalSpinControl(envOpts, false); ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~ /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_cpp/advanced_examples/ort/ort_priority_scheduling.cpp:998:57: warning: ignoring return value o f function declared with attribute warn_unused_result [-Wunused-result] Ort::GetApi().CreateEnvWithGlobalThreadPools(ORT_LOGGING_LEVEL_WARNING, "test", envOpts, &environment); ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ [ 85%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/tfl_main [ 88%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/dlr_main [ 92%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/ort_main [ 92%] Built target dlr_main [ 96%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/tfl_priority_scheduling [ 96%] Built target tfl_main [ 96%] Built target ort_main [100%] Linking CXX executable /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/bin/Release/ort_priority_scheduling [100%] Built target tfl_priority_scheduling [100%] Built target ort_priority_scheduling ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ source ./scripts/run_python_examples.sh X64 Architecture Running 3 Models - ['cl-tfl-mobilenet_v1_1.0_224', 'ss-tfl-deeplabv3_mnv2_ade20k_float', 'od-tfl-ssd_mobilenet_v2_300_float'] Running_Model : cl-tfl-mobilenet_v1_1.0_224 Downloading ../../../models/public/mobilenet_v1_1.0_224.tflite Running_Model : ss-tfl-deeplabv3_mnv2_ade20k_float Downloading ../../../models/public/deeplabv3_mnv2_ade20k_float.tflite Running_Model : od-tfl-ssd_mobilenet_v2_300_float Downloading ../../../models/public/ssd_mobilenet_v2_300_float.tflite /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/models/public/deeplabv3_mnv2_ade20k_float.tflite Preliminary number of subgraphs:1 , 81 nodes delegated out of 81 nodes Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal ************** Frame index 1 : Running float import ************* INFORMATION: [TIDL_ResizeLayer] ResizeBilinear_TIDL_0 Any resize ratio which is power of 2 and greater than 4 will be placed by combination of 4x4 resize layer and 2x2 resize layer. For example a 8x8 resize will be replaced by 4x4 resize followed by 2x2 resize. INFORMATION: [TIDL_ResizeLayer] ResizeBilinear_TIDL_1 Any resize ratio which is power of 2 and greater than 4 will be placed by combination of 4x4 resize layer and 2x2 resize layer. For example a 8x8 resize will be replaced by 4x4 resize followed by 2x2 resize. INFORMATION: [TIDL_ResizeLayer] ResizeBilinear Any resize ratio which is power of 2 and greater than 4 will be placed by combination of 4x4 resize layer and 2x2 resize layer. For example a 8x8 resize will be replaced by 4x4 resize followed by 2x2 resize. INFORMATION: [TIDL_ResizeLayer] decoder/ResizeBilinear Any resize ratio which is power of 2 and greater than 4 will be placed by combination of 4x4 resize layer and 2x 2 resize layer. For example a 8x8 resize will be replaced by 4x4 resize followed by 2x2 resize. INFORMATION: [TIDL_ResizeLayer] ResizeBilinear_1 Any resize ratio which is power of 2 and greater than 4 will be placed by combination of 4x4 resize layer and 2x2 resi ze layer. For example a 8x8 resize will be replaced by 4x4 resize followed by 2x2 resize. **************************************************** ** 5 WARNINGS 0 ERRORS ** **************************************************** The soft limit is 2048 The hard limit is 2048 MEM: Init ... !!! MEM: Init ... Done !!! 0.0s: VX_ZONE_INIT:Enabled 0.15s: VX_ZONE_ERROR:Enabled 0.22s: VX_ZONE_WARNING:Enabled 0.2368s: VX_ZONE_INIT:[tivxInit:184] Initialization Done !!! ************ Frame index 1 : Running float inference **************** Completed_Model : 2, Name : ss-tfl-deeplabv3_mnv2_ade20k_float , Total time : 28410.79, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_ss-tfl-deeplabv3_mnv2_ade20k_float_airshow.jpg MEM: Deinit ... !!! MEM: Alloc's: 26 alloc's of 684762304 bytes MEM: Free's : 26 free's of 684762304 bytes MEM: Open's : 0 allocs of 0 bytes MEM: Deinit ... Done !!! /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/models/public/mobilenet_v1_1.0_224.tflite Preliminary number of subgraphs:1 , 34 nodes delegated out of 34 nodes Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal ************** Frame index 1 : Running float import ************* **************************************************** ** ALL MODEL CHECK PASSED ** **************************************************** The soft limit is 2048 The hard limit is 2048 MEM: Init ... !!! MEM: Init ... Done !!! 0.0s: VX_ZONE_INIT:Enabled 0.19s: VX_ZONE_ERROR:Enabled 0.25s: VX_ZONE_WARNING:Enabled 0.2110s: VX_ZONE_INIT:[tivxInit:184] Initialization Done !!! ************ Frame index 1 : Running float inference **************** Completed_Model : 1, Name : cl-tfl-mobilenet_v1_1.0_224 , Total time : 1297.98, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_cl-tfl-mobilenet_v1_1.0_224_airshow.jpg MEM: Deinit ... !!! MEM: Alloc's: 26 alloc's of 106509608 bytes MEM: Free's : 26 free's of 106509608 bytes MEM: Open's : 0 allocs of 0 bytes MEM: Deinit ... Done !!! /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/models/public/ssd_mobilenet_v2_300_float.tflite Number of OD backbone nodes = 89 Size of odBackboneNodeIds = 89 Preliminary number of subgraphs:1 , 107 nodes delegated out of 107 nodes Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal ************** Frame index 1 : Running float import ************* **************************************************** ** ALL MODEL CHECK PASSED ** **************************************************** The soft limit is 2048 The hard limit is 2048 MEM: Init ... !!! MEM: Init ... Done !!! 0.0s: VX_ZONE_INIT:Enabled 0.3s: VX_ZONE_ERROR:Enabled 0.5s: VX_ZONE_WARNING:Enabled 0.1020s: VX_ZONE_INIT:[tivxInit:184] Initialization Done !!! ************ Frame index 1 : Running float inference **************** Completed_Model : 3, Name : od-tfl-ssd_mobilenet_v2_300_float , Total time : 4669.39, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_od-tfl-ssd_mobilenet_v2_300_float_airshow.jpg MEM: Deinit ... !!! MEM: Alloc's: 29 alloc's of 340181984 bytes MEM: Free's : 29 free's of 340181984 bytes MEM: Open's : 0 allocs of 0 bytes MEM: Deinit ... Done !!! Running 3 Models - ['cl-tfl-mobilenet_v1_1.0_224', 'ss-tfl-deeplabv3_mnv2_ade20k_float', 'od-tfl-ssd_mobilenet_v2_300_float'] Running_Model : cl-tfl-mobilenet_v1_1.0_224 Running_Model : ss-tfl-deeplabv3_mnv2_ade20k_float Running_Model : od-tfl-ssd_mobilenet_v2_300_float ******** WARNING ******* : Could not open ../../../model-artifacts//cl-tfl-mobilenet_v1_1.0_224///allowedNode.txt for reading... Entire model will run on ARM without a ny delegation to TIDL ! Number of subgraphs:1 , 0 nodes delegated out of 34 nodes ******** WARNING ******* : Could not open ../../../model-artifacts//ss-tfl-deeplabv3_mnv2_ade20k_float///allowedNode.txt for reading... Entire model will run on ARM wi thout any delegation to TIDL ! Number of subgraphs:1 , 0 nodes delegated out of 81 nodes ******** WARNING ******* : Could not open ../../../model-artifacts//od-tfl-ssd_mobilenet_v2_300_float///allowedNode.txt for reading... Entire model will run on ARM wit hout any delegation to TIDL ! Number of subgraphs:1 , 0 nodes delegated out of 107 nodes INFO: Created TensorFlow Lite XNNPACK delegate for CPU. , 0 0.584728 warplane, military plane ,, 1 0.232489 missile ,, 2 0.159124 projectile, missile ,, 3 0.021585 aircraft carrier, carrier, flattop, attack air craft carrier ,, 4 0.000915 submarine, pigboat, sub, U-boat , Saving image to ../../../output_images/ Completed_Model : 1, Name : cl-tfl-mobilenet_v1_1.0_224 , Total time : 15.98, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_cl-tfl-mobilenet_v1_1.0_224_airshow.jpg INFO: Created TensorFlow Lite XNNPACK delegate for CPU. INFO: Created TensorFlow Lite XNNPACK delegate for CPU. Saving image to ../../../output_images/ Completed_Model : 3, Name : od-tfl-ssd_mobilenet_v2_300_float , Total time : 65.79, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_od-tfl-ssd_mobilenet_v2_300_float_ADE_val_00001801.jpg Saving image to ../../../output_images/ Completed_Model : 2, Name : ss-tfl-deeplabv3_mnv2_ade20k_float , Total time : 156.67, Offload Time : 0.00 , DDR RW MBs : 18446744073709.55 , Output File : py_out_ss-tfl-deeplabv3_mnv2_ade20k_float_ADE_val_00001801.jpg Available execution providers : ['TIDLExecutionProvider', 'TIDLCompilationProvider', 'CPUExecutionProvider'] Running 4 Models - ['cl-ort-resnet18-v1', 'cl-ort-caffe_squeezenet_v1_1', 'ss-ort-deeplabv3lite_mobilenetv2', 'od-ort-ssd-lite_mobilenetv2_fpn'] Running_Model : cl-ort-resnet18-v1 Downloading ../../../models/public/resnet18_opset9.onnx Traceback (most recent call last): File "onnxrt_ep.py", line 281, in run_model(model, mIdx) File "onnxrt_ep.py", line 129, in run_model download_model(models_configs, model) File "/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/common_utils.py", line 199, in download_model onnxOpt.tidlOnnxModelOptimize(abs_path,abs_path, scale, mean) File "/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/scripts/osrt_model_tools/onnx_tools/onnx_model_opt.py", line 68, in tidlOnnxModelOptimize model = onnx.load_model(in_model_path) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 121, in load_model model = load_model_from_string(s, format=format) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 158, in load_model_from_string return _deserialize(s, ModelProto()) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 99, in _deserialize decoded = cast(Optional[int], proto.ParseFromString(s)) google.protobuf.message.DecodeError: Error parsing message with type 'onnx.ModelProto' Available execution providers : ['TIDLExecutionProvider', 'TIDLCompilationProvider', 'CPUExecutionProvider'] Running 4 Models - ['cl-ort-resnet18-v1', 'cl-ort-caffe_squeezenet_v1_1', 'ss-ort-deeplabv3lite_mobilenetv2', 'od-ort-ssd-lite_mobilenetv2_fpn'] Running_Model : cl-ort-resnet18-v1 Traceback (most recent call last): File "onnxrt_ep.py", line 281, in run_model(model, mIdx) File "onnxrt_ep.py", line 188, in run_model sess = rt.InferenceSession(config['model_path'] ,providers=EP_list, provider_options=[delegate_options, {}], sess_options=so) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 283, in __init__ self._create_inference_session(providers, provider_options) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 310, in _create_inference_session sess = C.InferenceSession(session_options, self._model_path, True, self._read_config_from_model) onnxruntime.capi.onnxruntime_pybind11_state.InvalidProtobuf: [ONNXRuntimeError] : 7 : INVALID_PROTOBUF : Load model from ../../../models/public/resnet18_opset9.onnx fa iled:Protobuf parsing failed. Downloading ../../../models/public/mobilenetv2-1.0.onnx Traceback (most recent call last): File "tvm_compilation_onnx_example.py", line 24, in download_model(models_configs, model_id) File "/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/common_utils.py", line 199, in download_model onnxOpt.tidlOnnxModelOptimize(abs_path,abs_path, scale, mean) File "/home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/scripts/osrt_model_tools/onnx_tools/onnx_model_opt.py", line 68, in tidlOnnxModelOptimize model = onnx.load_model(in_model_path) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 121, in load_model model = load_model_from_string(s, format=format) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 158, in load_model_from_string return _deserialize(s, ModelProto()) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 99, in _deserialize decoded = cast(Optional[int], proto.ParseFromString(s)) google.protobuf.message.DecodeError: Error parsing message with type 'onnx.ModelProto' Downloading ../../../models/public/inception_v3.tflite /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/models/public/inception_v3.tflite Generating subgraph boundary tensors for calibration... Building graph on host for tensor data collection... conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. Running graph on host for tensor data collection... Importing subgraph into TIDL... Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 16296.69 .... ..... ... .... ..... # 1 . .. T 16259.48 .... ..... ... .... ..... ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10735.66 .... ..... ... .... ..... # 1 . .. T 10737.59 .... ..... ... .... ..... ***************** Calibration iteration number 0 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10739.57 .... ..... ... .... ..... # 1 . .. T 10746.17 .... ..... ... .... ..... ***************** Calibration iteration number 1 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10814.22 .... ..... ... .... ..... # 1 . .. T 10785.53 .... ..... ... .... ..... ***************** Calibration iteration number 2 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10748.98 .... ..... ... .... ..... # 1 . .. T 10702.62 .... ..... ... .... ..... ***************** Calibration iteration number 3 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/tidl_import_subgraph0.txt.q unat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10749.14 .... ..... ... .... ..... # 1 . .. T 10731.21 .... ..... ... .... ..... ***************** Calibration iteration number 4 completed ************************ ------------------ Network Compiler Traces ----------------------------- NC running for device: 1 Running with OTF buffer optimizations Error : Error Code = Could not open /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3/tempDir/subgraph0_net/perfSimInfo.bin Rerunning network compiler for reshape ------------------ Network Compiler Traces ----------------------------- NC running for device: 1 Running with OTF buffer optimizations successful Memory allocation **************************************************** ** ALL MODEL CHECK PASSED ** **************************************************** TIDL import of 1 Relay IR subgraphs succeeded. TIDL artifacts are stored at ../../../model-artifacts/cl-dlr-tflite_inceptionnetv3 ../../../models/public/mobilenetv2-1.0.onnx Traceback (most recent call last): File "tvm_compilation_onnx_example.py", line 44, in onnx_model = onnx.load(model_path) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 121, in load_model model = load_model_from_string(s, format=format) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 158, in load_model_from_string return _deserialize(s, ModelProto()) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/__init__.py", line 99, in _deserialize decoded = cast(Optional[int], proto.ParseFromString(s)) google.protobuf.message.DecodeError: Error parsing message with type 'onnx.ModelProto' Generating subgraph boundary tensors for calibration... Building graph on host for tensor data collection... conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. conv2d NHWC layout is not optimized for x86 with autotvm. Running graph on host for tensor data collection... Importing subgraph into TIDL... Warning : Requested Output Data Convert Layer is not Added to the network, It is currently not Optimal ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 16300.13 .... ..... ... .... ..... # 1 . .. T 16352.09 .... ..... ... .... ..... ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10724.28 .... ..... ... .... ..... # 1 . .. T 10731.48 .... ..... ... .... ..... ***************** Calibration iteration number 0 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10747.59 .... ..... ... .... ..... # 1 . .. T 10725.32 .... ..... ... .... ..... ***************** Calibration iteration number 1 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10735.65 .... ..... ... .... ..... # 1 . .. T 10726.06 .... ..... ... .... ..... ***************** Calibration iteration number 2 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10759.03 .... ..... ... .... ..... # 1 . .. T 10732.85 .... ..... ... .... ..... ***************** Calibration iteration number 3 completed ************************ ~~~~~Running TIDL in PC emulation mode to collect Activations range for each layer~~~~~ Processing config file #0 : /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/tidl_import_subgraph 0.txt.qunat_stats_config.txt Freeing memory for user provided Net ----------------------- TIDL Process with REF_ONLY FLOW ------------------------ # 0 . .. T 10726.32 .... ..... ... .... ..... # 1 . .. T 10694.79 .... ..... ... .... ..... ***************** Calibration iteration number 4 completed ************************ ------------------ Network Compiler Traces ----------------------------- NC running for device: 1 Running with OTF buffer optimizations Error : Error Code = Could not open /home/ubuntu/Documents/edgeai-tidl-tools/edgeai-tidl-tools/model-artifacts/cl-dlr-tflite_inceptionnetv3_device/tempDir/subgraph0_net/perfSimInfo.bin Rerunning network compiler for reshape ------------------ Network Compiler Traces ----------------------------- NC running for device: 1 Running with OTF buffer optimizations successful Memory allocation **************************************************** ** ALL MODEL CHECK PASSED ** **************************************************** TIDL import of 1 Relay IR subgraphs succeeded. TIDL artifacts are stored at ../../../model-artifacts/cl-dlr-tflite_inceptionnetv3_device Traceback (most recent call last): File "tvm_compilation_mxnet_example.py", line 38, in from gluoncv import model_zoo File "/home/ubuntu/.local/lib/python3.6/site-packages/gluoncv/__init__.py", line 16, in _require_mxnet_version('1.4.0', '2.0.0') File "/home/ubuntu/.local/lib/python3.6/site-packages/gluoncv/check.py", line 6, in _require_mxnet_version import mxnet as mx File "/home/ubuntu/.local/lib/python3.6/site-packages/mxnet/__init__.py", line 24, in from .context import Context, current_context, cpu, gpu, cpu_pinned File "/home/ubuntu/.local/lib/python3.6/site-packages/mxnet/context.py", line 24, in from .base import classproperty, with_metaclass, _MXClassPropertyMetaClass File "/home/ubuntu/.local/lib/python3.6/site-packages/mxnet/base.py", line 213, in _LIB = _load_lib() File "/home/ubuntu/.local/lib/python3.6/site-packages/mxnet/base.py", line 204, in _load_lib lib = ctypes.CDLL(lib_path[0], ctypes.RTLD_LOCAL) File "/usr/lib/python3.6/ctypes/__init__.py", line 348, in __init__ self._handle = _dlopen(self._name, mode) OSError: libcudart.so.10.1: cannot open shared object file: No such file or directory Running Inference on Model - ../../../model-artifacts/cl-dlr-tflite_inceptionnetv3 2023-07-28 04:19:09,296 INFO Could not find libdlr.so in model artifact. Using dlr from /home/ubuntu/.local/lib/python3.6/site-packages/dlr/libdlr.so The soft limit is 2048 The hard limit is 2048 MEM: Init ... !!! MEM: Init ... Done !!! 0.0s: VX_ZONE_INIT:Enabled 0.9s: VX_ZONE_ERROR:Enabled 0.10s: VX_ZONE_WARNING:Enabled 0.1030s: VX_ZONE_INIT:[tivxInit:184] Initialization Done !!! Processing time in ms : 16075.1 , 0 0.509659 warplane, military plane ,, 1 0.327962 aircraft carrier, carrier, flattop, attack aircraft carrier ,, 2 0.023253 missile ,, 3 0.023253 project ile, missile ,, 4 0.002215 wing , Saving image to ../../../output_images/ Completed_Model : 1, Name : cl-dlr-tflite_inceptionnetv3 , Total time : 16075.15, Offload Time : 16075.15 , DDR RW MBs : 0, Output File : py_out_cl-dlr-tflite_inceptionnetv3_airshow.jpg MEM: Deinit ... !!! MEM: Alloc's: 26 alloc's of 125344384 bytes MEM: Free's : 26 free's of 125344384 bytes MEM: Open's : 0 allocs of 0 bytes MEM: Deinit ... Done !!! Running Inference on Model - ../../../model-artifacts/cl-dlr-onnx_mobilenetv2 2023-07-28 04:19:58,482 ERROR error in DLRModel instantiation model_path ../../../model-artifacts/cl-dlr-onnx_mobilenetv2 doesn't exist Traceback (most recent call last): File "/home/ubuntu/.local/lib/python3.6/site-packages/dlr/api.py", line 89, in __init__ self._impl = DLRModelImpl(model_path, dev_type, dev_id, error_log_file, use_default_dlr) File "/home/ubuntu/.local/lib/python3.6/site-packages/dlr/dlr_model.py", line 65, in __init__ raise ValueError("model_path %s doesn't exist" % model_path) ValueError: model_path ../../../model-artifacts/cl-dlr-onnx_mobilenetv2 doesn't exist Traceback (most recent call last): File "dlr_inference_example.py", line 211, in postprocess_for_onnx_mobilenetv2, 1) File "dlr_inference_example.py", line 164, in model_create_and_run model = DLRModel(model_dir, 'cpu') File "/home/ubuntu/.local/lib/python3.6/site-packages/dlr/api.py", line 92, in __init__ raise ex File "/home/ubuntu/.local/lib/python3.6/site-packages/dlr/api.py", line 89, in __init__ self._impl = DLRModelImpl(model_path, dev_type, dev_id, error_log_file, use_default_dlr) File "/home/ubuntu/.local/lib/python3.6/site-packages/dlr/dlr_model.py", line 65, in __init__ raise ValueError("model_path %s doesn't exist" % model_path) ValueError: model_path ../../../model-artifacts/cl-dlr-onnx_mobilenetv2 doesn't exist ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ ls bin docs lib onnx_1.7.0_x86_u18_bak requirements_pc.txt tflite_2.8_x86_u18 build entrypoint.sh libdlr.so opencv_4.2.0_x86_u18 run.sh tidl_tools dlr_1.10.0_x86_u18 examples LICENSE output_images scripts Dockerfile gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu model-artifacts README.md setup.sh dockers gcc-arm-9.2-2019.12-x86_64-aarch64-none-linux-gnu.tar.xz models requirements_pc_bak.txt test_data ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools$ cd examples/ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples$ ls cmake CMakeLists.txt jupyter_notebooks osrt_cpp osrt_python tidlrt_cpp ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples$ cd osrt_python/ort/ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ ls onnxrt_ep_gai.py onnxrt_ep.py ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ python3 onnxrt_ep.py -c Available execution providers : ['TIDLExecutionProvider', 'TIDLCompilationProvider', 'CPUExecutionProvider'] Running 4 Models - ['cl-ort-resnet18-v1', 'cl-ort-caffe_squeezenet_v1_1', 'ss-ort-deeplabv3lite_mobilenetv2', 'od-ort-ssd-lite_mobilenetv2_fpn'] Running_Model : cl-ort-resnet18-v1 Running shape inference on model ../../../models/public/resnet18_opset9.onnx Traceback (most recent call last): File "onnxrt_ep.py", line 281, in run_model(model, mIdx) File "onnxrt_ep.py", line 170, in run_model onnx.shape_inference.infer_shapes_path(config['model_path'], config['model_path']) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnx/shape_inference.py", line 60, in infer_shapes_path C.infer_shapes_path(model_path, output_path, check_type, strict_mode) onnx.onnx_cpp2py_export.checker.ValidationError: Unable to parse model from file:../../../models/public/resnet18_opset9.onnx. Please check if it is a valid protobuf file of model. ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ python3 onnxrt_ep.py Available execution providers : ['TIDLExecutionProvider', 'TIDLCompilationProvider', 'CPUExecutionProvider'] Running 4 Models - ['cl-ort-resnet18-v1', 'cl-ort-caffe_squeezenet_v1_1', 'ss-ort-deeplabv3lite_mobilenetv2', 'od-ort-ssd-lite_mobilenetv2_fpn'] Running_Model : cl-ort-resnet18-v1 Traceback (most recent call last): File "onnxrt_ep.py", line 281, in run_model(model, mIdx) File "onnxrt_ep.py", line 188, in run_model sess = rt.InferenceSession(config['model_path'] ,providers=EP_list, provider_options=[delegate_options, {}], sess_options=so) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 283, in __init__ self._create_inference_session(providers, provider_options) File "/home/ubuntu/.local/lib/python3.6/site-packages/onnxruntime/capi/onnxruntime_inference_collection.py", line 310, in _create_inference_session sess = C.InferenceSession(session_options, self._model_path, True, self._read_config_from_model) onnxruntime.capi.onnxruntime_pybind11_state.InvalidProtobuf: [ONNXRuntimeError] : 7 : INVALID_PROTOBUF : Load model from ../../../models/public/resnet18_opset9.onnx failed:Protobuf parsing failed. ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ pip3 show onnx Name: onnx Version: 1.9.0 Summary: Open Neural Network Exchange Home-page: https://github.com/onnx/onnx Author: ONNX Author-email: onnx-technical-discuss@lists.lfai.foundation License: Apache License v2.0 Location: /home/ubuntu/.local/lib/python3.6/site-packages Requires: numpy, protobuf, six, typing-extensions Required-by: caffe2onnx, onnx-simplifier, paddle2onnx ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ pip3 show onnxruntime WARNING: Package(s) not found: onnxruntime ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ pip3 list | grep onnxruntime onnxruntime-tidl 1.7.0 ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ pip3 show protobuf Name: protobuf Version: 3.19.4 Summary: Protocol Buffers Home-page: https://developers.google.com/protocol-buffers/ Author: Author-email: License: 3-Clause BSD License Location: /home/ubuntu/.local/lib/python3.6/site-packages Requires: Required-by: caffe2onnx, onnx, onnxruntime-tidl, paddle2onnx, tensorboard, tensorflow ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ python3 --version Python 3.6.9 ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$ uname -a Linux ubuntu-desktop 5.4.0-150-generic #167~18.04.1-Ubuntu SMP Wed May 24 00:51:42 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux ubuntu@ubuntu-desktop:~/Documents/edgeai-tidl-tools/edgeai-tidl-tools/examples/osrt_python/ort$