Start command ``` python -m vllm.entrypoints.openai.api_server --model baichuan-inc/Baichuan-13B-Chat --host 0.0.0.0 --port 8777 --trust-remote-code --dtype half ``` After about 12 hours of operation, the inference service stopped working GPU:V100 CUDA:11.4 Screenshot of the problem: 