TL

NVIDIA/TensorRT-LLM

TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in a performant way.

13.3k 2.3k +4/wk
GitHub
blackwell cuda llm-serving moe pytorch
Trend 3

Star & Fork Trend (30 data points)

Stars
Forks

Multi-Source Signals

Growth Velocity

NVIDIA/TensorRT-LLM has +4 stars this period . 7-day velocity: 0.2%.

Deep analysis is being generated for this repository.

Signal-backed technical analysis will be available soon.

No comparable projects found in the same topic categories.

Maintenance Activity 100

Last code push 0 days ago.

Community Engagement 85

Fork-to-star ratio: 17.0%. Active community forking and contributing.

Issue Burden 70

Issue data not yet available.

Growth Momentum 42

+4 stars this period — 0.03% growth rate.

License Clarity 30

No clear license detected — proceed with caution.

Risk scores are computed from real-time repository data. Higher scores indicate healthier metrics.