You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I reviewed you guide to train a TF2 object detection model and wonder why you're using two different datatypes for the final inference.
First you stick with int8 and then you declare uint8 for the input_tensor. The second weird thing is, the final use of float32.
I assume the model get feeded with uint8 (should be int8) for faster inference and the output_tensor should be as accurate as possible? Didn't you build a bottleneck with that?
Logically would be using int8 for the input_tensor and the output_tensor or not.
# For full integer quantization, though supported types defaults to int8 only, we explicitly declare it for clarity.converter.target_spec.supported_types= [tf.int8]
# These set the input tensors to uint8 and output tensors to float32converter.inference_input_type=tf.uint8converter.inference_output_type=tf.float32
The text was updated successfully, but these errors were encountered:
Hey,
I reviewed you guide to train a TF2 object detection model and wonder why you're using two different datatypes for the final inference.
First you stick with
int8
and then you declareuint8
for the input_tensor. The second weird thing is, the final use offloat32
.I assume the model get feeded with
uint8
(should beint8
) for faster inference and the output_tensor should be as accurate as possible? Didn't you build a bottleneck with that?Logically would be using
int8
for the input_tensor and the output_tensor or not.The text was updated successfully, but these errors were encountered: