AI
Agent Directory
NewsBlogBrowseBenchmarkSubmitFAQAbout
AI
Agent Directory

The home for AI agents, frameworks, and tools. Discover what's next.

Explore

  • Browse All
  • News
  • Submit Listing
  • FAQ
  • API

Company

  • About
  • Contact
  • Privacy
  • Terms

Community

  • X
  • GitHub
  • LinkedIn
© 2026 AI Agent Directory
Home / Listings / TensorRT-LLM

TensorRT-LLM

A high-performance inference engine for large language models.

frameworkOpen Source
FrameworksInfrastructure

TensorRT-LLM is a high-performance inference engine designed for running large language models efficiently. It leverages NVIDIA's TensorRT technology to optimize model performance, enabling faster inference times and lower latency. TensorRT-LLM is particularly beneficial for applications that require real-time processing of language data, such as chatbots and virtual assistants. Its architecture is optimized for NVIDIA GPUs, making it a powerful solution for developers looking to deploy large models in production environments.

Visit Website →GitHubDocs
1 views0 clicksAdded 3/14/2026

Reviews

No reviews yet. Be the first!

Loading reviews...

Advertise Here

Reach AI developers and builders

From $49/mo →