You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+15
Original file line number
Diff line number
Diff line change
@@ -180,6 +180,21 @@ key: "ENABLE_WEIGHT_SHARING"
180
180
}
181
181
```
182
182
183
+
*`ENABLE_DETERMINISTIC_ALGORITHMS`: Boolean flag to enable deterministic algorithm selection for TorchScript models. By default, deterministic algorithms are disabled.
184
+
185
+
When this flag is set to `true`, Triton will configure the PyTorch backend to use only deterministic algorithm implementations. This ensures that model outputs are reproducible across runs, at the cost of potential performance degradation. If any operation does not have a deterministic version, an error will be raised.
186
+
187
+
The section of model config file specifying this parameter will look like:
188
+
189
+
```
190
+
parameters: {
191
+
key: "ENABLE_DETERMINISTIC_ALGORITHMS"
192
+
value: {
193
+
string_value: "true"
194
+
}
195
+
}
196
+
```
197
+
183
198
*`ENABLE_CACHE_CLEANING`: Boolean flag to enable CUDA cache cleaning after each model execution.
184
199
If not specified, cache cleaning is disabled. This flag has no effect if model is on CPU.
185
200
Setting this flag to true will negatively impact the performance due to additional CUDA cache
0 commit comments