[1;36m(EngineCore_DP0 pid=2738693)[0;0m ERROR 10-10 10:43:10 [core.py:710] File "/aisoft/conda/env/vllm2/lib/python3.10/concurrent/futures/_base.py", line 458, in result
[1;36m(EngineCore_DP0 pid=2738693)[0;0m ERROR 10-10 10:43:10 [core.py:710] File "/aisoft/conda/env/vllm2/lib/python3.10/concurrent/futures/thread.py", line 58, in run
[1;36m(EngineCore_DP0 pid=2738693)[0;0m ERROR 10-10 10:43:10 [core.py:710] result = self.fn(*self.args, **self.kwargs)
[1;36m(EngineCore_DP0 pid=2738693)[0;0m ERROR 10-10 10:43:10 [core.py:710] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/__init__.py", line 128, in _async_create_grammar
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [async_llm.py:480] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/core_client.py", line 846, in get_output_async
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [async_llm.py:480] raise self._format_exception(outputs) from None
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [async_llm.py:480] vllm.v1.engine.exceptions.EngineDeadError: EngineCore encountered an issue. See stack trace (above) for the root cause.
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/multiprocessing/process.py", line 314, in _bootstrap
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 712, in run_engine_core
[1;36m(EngineCore_DP0 pid=2738693)[0;0m raise e
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 701, in run_engine_core
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/core.py", line 754, in _process_engine_step
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/core/sched/scheduler.py", line 359, in schedule
[1;36m(EngineCore_DP0 pid=2738693)[0;0m if structured_output_req and structured_output_req.grammar:
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/request.py", line 45, in grammar
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/request.py", line 33, in _check_grammar_completion
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/concurrent/futures/thread.py", line 58, in run
[1;36m(EngineCore_DP0 pid=2738693)[0;0m result = self.fn(*self.args, **self.kwargs)
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/__init__.py", line 128, in _async_create_grammar
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/functools.py", line 981, in __get__
[1;36m(EngineCore_DP0 pid=2738693)[0;0m val = self.func(instance)
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/request.py", line 58, in structured_output_key
[1;36m(EngineCore_DP0 pid=2738693)[0;0m File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/structured_output/request.py", line 86, in get_structured_output_key
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/entrypoints/openai/serving_chat.py", line 574, in chat_completion_stream_generator
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] async for res in result_generator:
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/async_llm.py", line 387, in generate
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] out = q.get_nowait() or await q.get()
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/output_processor.py", line 59, in get
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] File "/aisoft/conda/env/vllm2/lib/python3.10/site-packages/vllm/v1/engine/core_client.py", line 846, in get_output_async
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] raise self._format_exception(outputs) from None
[1;36m(APIServer pid=2738423)[0;0m ERROR 10-10 10:43:10 [serving_chat.py:1145] vllm.v1.engine.exceptions.EngineDeadError: EngineCore encountered an issue. See stack trace (above) for the root cause.
[rank0]:[W1010 10:43:10.666166970 ProcessGroupNCCL.cpp:1538] Warning: WARNING: destroy_process_group() was not called before program exit, which can leak resources. For more info, please see https://pytorch.org/docs/stable/distributed.html#shutdown (function operator())
[1;36m(APIServer pid=2738423)[0;0m INFO: Shutting down
[1;36m(APIServer pid=2738423)[0;0m INFO: Waiting for application shutdown.