Open
Description
System Info
Nvidia rtx 3090 ti
nvcr.io/nvidia/tritonserver:25.05-trtllm-python-py3
Who can help?
Information
- The official example scripts
- My own modified scripts
Tasks
- An officially supported task in the
examples
folder (such as GLUE/SQuAD, ...) - My own task or dataset (give details below)
Reproduction
Steps to reproduce the behaviour:
- take any tensorrt_llm compiled plan
- delete from config.pbtxt parameters
tokenizer_dir
,xgrammar_tokenizer_info_path
orguided_decoding_backend
Expected behavior
tensorrtllm_backend
should start normally
actual behavior
tensorrtllm_backend
crashes with a message that a parameter is missing, although it's not used.
additional notes
TensorRT-LLM/triton_backend/inflight_batcher_llm/src/model_instance_state.cc
Lines 410 to 453 in 7b210ae