Skip to content

Pull requests: triton-inference-server/tensorrtllm_backend

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

Replace subprocess.Popen with subprocess.run triaged Issue has been triaged by maintainers
#452 opened May 14, 2024 by rlempka Loading…
add speculative decoding example
#432 opened Apr 24, 2024 by XiaobingSuper Loading…
Fixed Whitespace Error in Streaming mode
#423 opened Apr 19, 2024 by enochlev Loading…
Update end_to_end_test.py
#409 opened Apr 14, 2024 by r0cketdyne Loading…
fix: add foreground argument
#343 opened Feb 21, 2024 by pfldy2850 Loading…
Expose verbose as pram in launch triton script
#295 opened Jan 12, 2024 by ekagra-ranjan Loading…
Add example of tensorrt-llm usage
#225 opened Dec 15, 2023 by Pernekhan Loading…
Wrap long command-lines in README.md
#134 opened Nov 15, 2023 by wangkuiyi Loading…
draft pr about non-streaming output
#95 opened Nov 3, 2023 by BasicCoder Loading…
ProTip! Exclude everything labeled bug with -label:bug.