AI
Agent Directory
NewsBlogBrowseBenchmarkSubmitFAQAbout
AI
Agent Directory

The home for AI agents, frameworks, and tools. Discover what's next.

Explore

  • Browse All
  • News
  • Submit Listing
  • FAQ
  • API

Company

  • About
  • Contact
  • Privacy
  • Terms

Community

  • X
  • GitHub
  • LinkedIn
© 2026 AI Agent Directory
Home / Listings / vLLM
vLLM logo

vLLM

⭐ Featured

High-throughput open-source LLM inference engine

toolFree
Infrastructure#self-hosted#open-source#inference#performance

vLLM is the most popular open-source LLM inference engine. It implements PagedAttention for efficient memory management, achieving 2-4x higher throughput than naive serving. vLLM supports continuous batching, tensor parallelism, speculative decoding, and serves an OpenAI-compatible API. It's the standard engine behind most self-hosted LLM deployments.

Visit Website →GitHub
4 views0 clicksAdded 3/14/2026

Reviews

No reviews yet. Be the first!

Loading reviews...

Advertise Here

Reach AI developers and builders

From $49/mo →