STARTING! Log parameter values verbosely: [1] Min num runs: [50] Min runs duration (seconds): [1] Max runs duration (seconds): [150] Inter-run delay (seconds): [-1] Number of prorated runs per second: [-1] Num threads: [4] Use caching: [0] Benchmark name: [] Output prefix: [] Min warmup runs: [1] Min warmup runs duration (seconds): [0.5] Run w/o invoking kernels: [0] Report the peak memory footprint: [0] Memory footprint check interval (ms): [50] Graph: [/Users/luohanjie/Workspace/Vision/my_slam/data/models/model_opt.tflite] Input layers: [] Input shapes: [] Input value ranges: [] Input value files: [] Allow fp16: [0] Require full delegation: [0] Enable op profiling: [0] Max initial profiling buffer entries: [1024] Allow dynamic increase on profiling buffer entries: [0] CSV File to export profiling data to: [] Print pre-invoke interpreter state: [0] Print post-invoke interpreter state: [0] Release dynamic tensor memory: [0] Use dynamic tensor for large tensors: [0] print out all supported flags: [0] #threads used for CPU inference: [4] Max number of delegated partitions: [0] Min nodes per partition: [0] Directory for delegate serialization: [] Model-specific token/key for delegate serialization.: [] Use xnnpack: [0] External delegate path: [] External delegate options: [] Use gpu: [0] Allow lower precision in gpu: [1] Enable running quant models in gpu: [1] Prefer maximizing the throughput in gpu: [0] GPU backend: [] Loaded model /Users/luohanjie/Workspace/Vision/my_slam/data/models/model_opt.tflite INFO: Created TensorFlow Lite XNNPACK delegate for CPU. The input model file size (MB): 66.3383 Initialized session in 41.498ms. Running benchmark for at least 1 iterations and at least 0.5 seconds but terminate if exceeding 150 seconds. count=13 first=43827 curr=38759 min=38662 max=45293 avg=39973.3 std=1998
Running benchmark for at least 50 iterations and at least 1 seconds but terminate if exceeding 150 seconds. count=50 first=39240 curr=38747 min=38470 max=40766 avg=39654.3 std=635
Inference timings in us: Init: 41498, First inference: 43827, Warmup (avg): 39973.3, Inference (avg): 39654.3
STARTING! Log parameter values verbosely: [1] Min num runs: [50] Min runs duration (seconds): [1] Max runs duration (seconds): [150] Inter-run delay (seconds): [-1] Number of prorated runs per second: [-1] Num threads: [4] Use caching: [0] Benchmark name: [] Output prefix: [] Min warmup runs: [1] Min warmup runs duration (seconds): [0.5] Run w/o invoking kernels: [0] Report the peak memory footprint: [0] Memory footprint check interval (ms): [50] Graph: [/Users/luohanjie/Workspace/Vision/my_slam/data/models/model_opt.tflite] Input layers: [] Input shapes: [] Input value ranges: [] Input value files: [] Allow fp16: [0] Require full delegation: [0] Enable op profiling: [0] Max initial profiling buffer entries: [1024] Allow dynamic increase on profiling buffer entries: [0] CSV File to export profiling data to: [] Print pre-invoke interpreter state: [0] Print post-invoke interpreter state: [0] Release dynamic tensor memory: [0] Use dynamic tensor for large tensors: [0] print out all supported flags: [0] #threads used for CPU inference: [4] Max number of delegated partitions: [0] Min nodes per partition: [0] Directory for delegate serialization: [] Model-specific token/key for delegate serialization.: [] Use xnnpack: [0] External delegate path: [] External delegate options: [] Use gpu: [1] Allow lower precision in gpu: [1] Enable running quant models in gpu: [1] Prefer maximizing the throughput in gpu: [0] GPU backend: [] Loaded model /Users/luohanjie/Workspace/Vision/my_slam/data/models/model_opt.tflite INFO: Created TensorFlow Lite delegate for GPU. GPU delegate created. INFO: Initialized OpenCL-based API. INFO: Created 1 GPU delegate kernels. Explicitly applied GPU delegate, and the model graph will be completely executed by the delegate. The input model file size (MB): 66.3383 Initialized session in 129.521ms. Running benchmark for at least 1 iterations and at least 0.5 seconds but terminate if exceeding 150 seconds. count=40 first=40053 curr=11752 min=11744 max=40053 avg=12579.9 std=4400
Running benchmark for at least 50 iterations and at least 1 seconds but terminate if exceeding 150 seconds. count=85 first=11880 curr=11836 min=11567 max=12276 avg=11839.5 std=93
Inference timings in us: Init: 129521, First inference: 40053, Warmup (avg): 12579.9, Inference (avg): 11839.5