Skip to content

Pull requests: NVIDIA/TensorRT-LLM

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

@coderabbitai title
#6251 opened Jul 22, 2025 by venkywonka Loading…
Improve TransferAgentTest.SyncMessage
#6250 opened Jul 22, 2025 by bo-nv Loading…
update disagg slurm scripts
#6248 opened Jul 22, 2025 by lingjiew Loading…
doc: update known issues
#6247 opened Jul 22, 2025 by QiJune Loading…
[feat] Support NVFP4 KV Cache
#6244 opened Jul 22, 2025 by Tom-Zheng Loading…
Chore: remove duplicate should_stop_processing check
#6242 opened Jul 22, 2025 by QiJune Loading…
[PERF] Don't use hmac encryption for loopback interfaces Community want to contribute PRs initiated from Community
#6241 opened Jul 22, 2025 by vadiklyutiy Loading…
fix: WAR for nvbug_5398806
#6239 opened Jul 21, 2025 by hchings Loading…
[Issue 6193] Fix gemma3vl weight loader
#6233 opened Jul 21, 2025 by johncalesp Loading…
Auto-enable ngram with concurrency <= 32.
#6232 opened Jul 21, 2025 by SimengLiu-nv Loading…
Remove input_sf swizzle for module WideEPMoE
#6231 opened Jul 21, 2025 by StudyingShao Loading…
Waive flaky tests
#6229 opened Jul 21, 2025 by Tabrizian Loading…
[fix] Allow custom model config for Kimi-K2 Community want to contribute PRs initiated from Community
#6228 opened Jul 21, 2025 by meenchen Loading…
Change the all-reduce strategy to NCCL
#6226 opened Jul 21, 2025 by nzmora-nvidia Loading…
Bump version to 0.21.1
#6225 opened Jul 21, 2025 by yuanjingx87 Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.