-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Open
Labels
InvestigatingLLM API<NV>High-level LLM Python API & tools (e.g., trtllm-llmapi-launch) for TRTLLM inference/workflows.<NV>High-level LLM Python API & tools (e.g., trtllm-llmapi-launch) for TRTLLM inference/workflows.questionFurther information is requestedFurther information is requestedstaletriagedIssue has been triaged by maintainersIssue has been triaged by maintainerswaiting for feedback
Description
Trying out T5 with python backend. https://github.com/NVIDIA/TensorRT-LLM/blob/main/examples/enc_dec/run.py#L484
I see SamplingConfig has output_log_probs https://github.com/NVIDIA/TensorRT-LLM/blob/main/tensorrt_llm/runtime/generation.py#L355.
But in the return dict does not have the log probabilities https://github.com/NVIDIA/TensorRT-LLM/blob/main/tensorrt_llm/runtime/generation.py#L2515.
Is there any other way to get the log probabilities?
Metadata
Metadata
Assignees
Labels
InvestigatingLLM API<NV>High-level LLM Python API & tools (e.g., trtllm-llmapi-launch) for TRTLLM inference/workflows.<NV>High-level LLM Python API & tools (e.g., trtllm-llmapi-launch) for TRTLLM inference/workflows.questionFurther information is requestedFurther information is requestedstaletriagedIssue has been triaged by maintainersIssue has been triaged by maintainerswaiting for feedback