-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Triton Crash with Signal 11 while using python backend #7400
Comments
I am getting a very similar problem, however I am not sure if it is the exact same error. I also have a python decupled backend. After starting tritonserver I run stress testing which sends a lot of requests to the tritonserver. Within the first 10 minutes of testing I quite consistently get this error which completely crushes my tritonserver. Unfortunately I have a custom build of tritonserver based on 24.05, so I don't know how relevant the information is. However, I did not have this problem on version 23.10
I assume the error occurs because of this check, however I have no clue why this is the case: server/src/grpc/infer_handler.h Lines 183 to 192 in c61d993
|
Description
After using the Python vllm backend, Triton crashed with signal 11. The model had been loaded and preheated for some time before the crash occurred.
Triton Information
What version of Triton are you using?
Are you using the Triton container or did you build it yourself?
Yes
trace info:
The text was updated successfully, but these errors were encountered: