You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"dla_core": 0, # (DLA only) Target dla core id to run engine
283
287
"allow_gpu_fallback": false, # (DLA only) Allow layers unsupported on DLA to run on GPU
284
288
},
285
-
"op_precision": torch.half, # Operating precision set to FP16
286
-
# List of datatypes that should be configured for each input. Supported options torch.{float|half|int8|int32|bool}.
289
+
"enabled_precisions": {torch.half}, # Operating precision set to FP16
290
+
"sparse_weights": Enable sparsity for convolution and fully connected layers.
287
291
"disable_tf32": False, # Force FP32 layers to use traditional as FP32 format vs the default behavior of rounding the inputs to 10-bit mantissas before multiplying, but accumulates the sum using 23-bit mantissas
"disable_tf32": False, # Force FP32 layers to use traditional as FP32 format vs the default behavior of rounding the inputs to 10-bit mantissas before multiplying, but accumulates the sum using 23-bit mantissas
48
+
"sparse_weights": Enable sparsity for convolution and fully connected layers.
47
49
"refit": false, # enable refit
48
50
"debug": false, # enable debuggable engine
49
51
"strict_types": false, # kernels should strictly run in operating precision
@@ -113,6 +115,7 @@ def convert_method_to_trt_engine(module: torch.jit.ScriptModule, method_name: st
# List of datatypes that should be configured for each input. Supported options torch.{float|half|int8|int32|bool}.
115
117
"disable_tf32": False, # Force FP32 layers to use traditional as FP32 format vs the default behavior of rounding the inputs to 10-bit mantissas before multiplying, but accumulates the sum using 23-bit mantissas
118
+
"sparse_weights": Enable sparsity for convolution and fully connected layers.
116
119
"refit": false, # enable refit
117
120
"debug": false, # enable debuggable engine
118
121
"strict_types": false, # kernels should strictly run in operating precision
0 commit comments