-
Notifications
You must be signed in to change notification settings - Fork 365
✨[Feature] Setting the input data type of models, such as INT32, is not supported #388
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@narendasan I agree we need to provide an interface to provide input data types of tensors. Also TRT supports INT32 input datatype. When I have two inputs (float32, int32) to the graph, TRT worked fine. I did not |
I think those inputs (float32, int32) cases are cases where the int32 are shape tensors typically. I am not sure there is actual input data in int32 (can be completely wrong as well). |
Oh yes there is actual input data in INT32 (INT64 even) for BERT and other transformers. |
Comment added by Nick Comly in Aha! - View If Input provided: Use specified precision Else: use engine precision (INT8 uses FP32 input) |
Is your feature request related to a problem? Please describe.
When i tried to use TRTorch to convert some PyTorch models which input's data type is INT32 such as BERT , it failed in
model.forward(inputs)
process. The error message is below. It seems that TRT Model's input data type is related to op_precision.Describe the solution you'd like
I would like to add a corresponding interface to specify the type of input data.
Describe alternatives you've considered
Additional context
When executing trt_mod.forward(), a bug appears

The text was updated successfully, but these errors were encountered: