Replies: 1 comment 1 reply
-
Solved. Actually, argument, responsible for this logic is
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I'm using Triton Server to start a ONNX model, that has pretty big batch. My goal is to run it, no matter how much time it will execute the batch only using CPU.
Now I'm stuck on the default Request timeout (30 secs) after which Triton aborts the request:
I've only found that doc: https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/protocol/extension_schedule_policy.html
But I don't need a dynamic batcher, is there any argument of model_configuration for a default batcher?
My model_configuration:
Triton server container version: "22.10"
UPD: Adding lines below to model config doesn't change timeout (it's still 30 secs)
Beta Was this translation helpful? Give feedback.
All reactions