root@6530a5cb9657:/home/root/examples/osrt_python/advanced_examples/unit_tests_validation/ort# python3 ./onnxrt_ep.py -c -m fast_bev_bilinear ['fast_bev_bilinear'] Available execution providers : ['TIDLExecutionProvider', 'TIDLCompilationProvider', 'CPUExecutionProvider'] Running 1 Models - ['fast_bev_bilinear'] Running_Model : fast_bev_bilinear Running shape inference on model ../unit_test_models/fast_bev_bilinear.onnx tidl_tools_path = /home/root/tidl_tools/ artifacts_folder = ../model-artifacts//fast_bev_bilinear/artifacts tidl_tensor_bits = 8 debug_level = 4 num_tidl_subgraphs = 16 num_tidl_subgraph_max_node = 0 enable_rt_multi_subgraph_support = 0 tidl_denylist = tidl_denylist_layer_name = tidl_denylist_layer_type = tidl_allowlist_layer_name = model_type = tidl_calibration_accuracy_level = 7 tidl_calibration_options:num_frames_calibration = 1 tidl_calibration_options:bias_calibration_iterations = 1 mixed_precision_factor = -1.000000 model_group_id = 0 power_of_2_quantization = 2 ONNX QDQ Enabled = 0 enable_high_resolution_optimization = 0 pre_batchnorm_fold = 1 add_data_convert_ops = 3 output_feature_16bit_names_list = m_params_16bit_names_list = m_single_core_layers_names_list = Inference mode = 0 Number of cores = 1 reserved_compile_constraints_flag = 83886080 partial_init_during_compile = 0 packetize_mode = 0 ti_internal_reserved_1 = ========================= [Model Compilation Started] ========================= Model compilation will perform the following stages: 1. Parsing 2. Graph Optimization 3. Quantization & Calibration 4. Memory Planning ============================== [Version Summary] ============================== ------------------------------------------------------------------------------- | TIDL Tools Version | 10_01_00_01 | ------------------------------------------------------------------------------- | C7x Firmware Version | 10_01_00_01 | ------------------------------------------------------------------------------- | Runtime Version | 1.15.0 | ------------------------------------------------------------------------------- | Model Opset Version | 16 | ------------------------------------------------------------------------------- NOTE: The runtime version here specifies ONNXRT_VERSION+TIDL_VERSION Ex: 1.14.0+1000XXXX -> ONNXRT 1.14.0 and a TIDL_VERSION 10.00.XX.XX ============================== [Parsing Started] ============================== [TIDL Import] [PARSER] WARNING: Network not identified as Object Detection network : (1) Ignore if network is not Object Detection network (2) If network is Object Detection network, please specify "model_type":"OD" as part of OSRT compilation options [TIDL Import] [PARSER] SUPPORTED: Layers type supported by TIDL --- layer type - GridSample, Node name - /bev_encoder/GridSample -- [tidl_onnxRtImport_core.cpp, 587] ------------------------- Subgraph Information Summary ------------------------- ------------------------------------------------------------------------------- | Core | No. of Nodes | Number of Subgraphs | ------------------------------------------------------------------------------- | C7x | 1 | 1 | | CPU | 0 | x | ------------------------------------------------------------------------------- Running Runtimes GraphViz - /home/root/tidl_tools//tidl_graphVisualiser_runtimes.out ../model-artifacts//fast_bev_bilinear/artifacts/allowedNode.txt ../model-artifacts//fast_bev_bilinear/artifacts/tempDir/graphvizInfo.txt ../model-artifacts//fast_bev_bilinear/artifacts/tempDir/runtimes_visualization.svg ============================= [Parsing Completed] ============================= TIDL_createStateImportFunc Started: Compute on node : TIDLExecutionProvider_TIDL_0_0 0, GridSample, 2, 1, /bev_encoder/Reshape_output_0, /bev_encoder/GridSample_output_0 Input tensor name - /bev_encoder/Reshape_output_0 Input tensor name - /bev_encoder/Unsqueeze_1_output_0 Output tensor name - /bev_encoder/GridSample_output_0 In TIDL_onnxRtImportInit subgraph_name=subgraph_0 Layer 0, subgraph id subgraph_0, name=/bev_encoder/GridSample_output_0 Layer 1, subgraph id subgraph_0, name=/bev_encoder/Reshape_output_0 Layer 2, subgraph id subgraph_0, name=/bev_encoder/Unsqueeze_1_output_0 ==================== [Optimization for subgraph_0 Started] ==================== In TIDL_runtimesOptimizeNet: LayerIndex = 4, dataIndex = 3 ----------------------------- Optimization Summary ----------------------------- ------------------------------------------------------------------------------- | Layer | Nodes before optimization | Nodes after optimization | ------------------------------------------------------------------------------- | TIDL_GridSampleLayer | 1 | 1 | ------------------------------------------------------------------------------- =================== [Optimization for subgraph_0 Completed] =================== In TIDL_runtimesPostProcessNet ************ in TIDL_subgraphRtCreate ************ The soft limit is 10240 The hard limit is 10240 MEM: Init ... !!! MEM: Init ... Done !!! 0.0s: VX_ZONE_INIT:Enabled 0.5s: VX_ZONE_ERROR:Enabled 0.9s: VX_ZONE_WARNING:Enabled 0.2218s: VX_ZONE_INIT:[tivxInit:190] Initialization Done !!! ************ TIDL_subgraphRtCreate done ************ ============= [Quantization & Calibration for subgraph_0 Started] ============= ******* In TIDL_subgraphRtInvoke ******** 0 1.00000 -4.85212 4.28586 6 2 1.00000 -4.85212 4.28586 6 1 1.00000 -4.85212 4.28586 6 3 1.00000 -4.85212 4.28586 6 4 1.00000 -4.52524 3.91146 6 5 1.00000 -4.52524 3.91146 6 Layer, Layer Cycles,kernelOnlyCycles, coreLoopCycles,LayerSetupCycles,dmaPipeupCycles, dmaPipeDownCycles, PrefetchCycles,copyKerCoeffCycles,LayerDeinitCycles,LastBlockCycles, paddingTrigger, paddingWait,LayerWithoutPad,LayerHandleCopy, BackupCycles, RestoreCycles,Multic7xContextCopyCycles, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 4, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 5, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, Sum of Layer Cycles 0 Sub Graph Stats 160.000000 139880.000000 10241.000000 ******* TIDL_subgraphRtInvoke done ******** In TIDL_runtimesPostProcessNet -------- Running Calibration in Float Mode to Collect Tensor Statistics -------- [=============================================================================] 100 % ------------------ Fixed-point Calibration Iteration [1 / 1]: ------------------ [=============================================================================] 100 % ==================== [Quantization & Calibration Completed] ==================== ========================== [Memory Planning Started] ========================== ------------------------- Network Compiler Traces ------------------------------ Successful Memory Allocation Successful Workload Creation ========================= [Memory Planning Completed] ========================= ======================== Subgraph Compiled Successfully ======================== Completed model - fast_bev_bilinear.onnx Name : fast_bev_bilinear , Total time : 1978.71, Offload Time : 139.88 , DDR RW MBs : 0 ************ in TIDL_subgraphRtDelete ************ MEM: Deinit ... !!! MEM: Alloc's: 28 alloc's of 353587469 bytes MEM: Free's : 28 free's of 353587469 bytes MEM: Open's : 0 allocs of 0 bytes MEM: Deinit ... Done !!!