Skip to content

Text Generation Inference (TGI)

⭐ 10.8k stars
Repositoryhuggingface/text-generation-inference
Categoryinfra
Difficultyadvanced
Statusactive
Tagsserving inference huggingface rust
Websitehttps://huggingface.co/docs/text-generation-inference

Review

Hugging Face's production-ready inference server for LLMs. Optimized with continuous batching, tensor parallelism, and quantization support. Best for Hugging Face ecosystem users deploying models in production.

Use Cases

  • model-serving
  • inference
  • deployment

Curated with care for the AI developer community