You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
when I run the code like this: options = onnxruntime.SessionOptions() options.enable_profiling=True session = onnxruntime.InferenceSession(model_path, providers=providers, sess_options=options) session.run_with_iobinding(io_binding) profile_file = session.end_profiling()
I got a json profile file, and visualize it with edge://tracing/, got the result as follows:
I have two questions here:
why the first conv cost so much? Is the time of loading data from host to device included?
except the first conv, add the latency of other ops, the result is significantly greater than the time I got in my time test code below. So how can I get the exact run time of every single op?
The text was updated successfully, but these errors were encountered:
This first Conv need run all possible kernels to find the fastest kernel for a given input shape. It will take longer than the remaining inference run of the same shape input.
when I run the code like this:
options = onnxruntime.SessionOptions()
options.enable_profiling=True
session = onnxruntime.InferenceSession(model_path, providers=providers, sess_options=options)
session.run_with_iobinding(io_binding)
profile_file = session.end_profiling()
I got a json profile file, and visualize it with
edge://tracing/
, got the result as follows:I have two questions here:
The text was updated successfully, but these errors were encountered: