diff --git a/sit4onnx/__init__.py b/sit4onnx/__init__.py index 5155e11..959f50e 100644 --- a/sit4onnx/__init__.py +++ b/sit4onnx/__init__.py @@ -1,3 +1,3 @@ from sit4onnx.onnx_inference_test import inference, main -__version__ = '1.0.6' +__version__ = '1.0.7' diff --git a/sit4onnx/onnx_inference_test.py b/sit4onnx/onnx_inference_test.py index 19e1c7c..a62efb0 100644 --- a/sit4onnx/onnx_inference_test.py +++ b/sit4onnx/onnx_inference_test.py @@ -36,8 +36,15 @@ class Color: ONNX_DTYPES_TO_NUMPY_DTYPES: dict = { + f'{onnx.TensorProto.FLOAT16}': np.float16, f'{onnx.TensorProto.FLOAT}': np.float32, f'{onnx.TensorProto.DOUBLE}': np.float64, + f'{onnx.TensorProto.UINT8}': np.uint8, + f'{onnx.TensorProto.UINT16}': np.uint16, + f'{onnx.TensorProto.UINT32}': np.uint32, + f'{onnx.TensorProto.UINT64}': np.uint64, + f'{onnx.TensorProto.INT8}': np.int8, + f'{onnx.TensorProto.INT16}': np.int16, f'{onnx.TensorProto.INT32}': np.int32, f'{onnx.TensorProto.INT64}': np.int64, } @@ -225,6 +232,7 @@ def inference( session_option = onnxruntime.SessionOptions() session_option.log_severity_level = 4 session_option.intra_op_num_threads = intra_op_num_threads + if sub_info: if onnx_execution_provider in ['openvino_cpu', 'openvino_gpu']: session_option.graph_optimization_level = onnxruntime.GraphOptimizationLevel.ORT_DISABLE_ALL