Infrastructure
Deployment, observability, evaluation, and ops
Ollama
⭐ 168.6kThe easiest way to run open-source LLMs locally.
vLLM
⭐ 76.2kHigh-throughput LLM inference engine with PagedAttention for efficient GPU memory management.
Qdrant
⭐ 30.3kHigh-performance vector database written in Rust with advanced filtering capabilities.
ChromaDB
⭐ 27.4kSimple open-source vector database for AI applications.
LangFuse
⭐ 24.8kOpen-source LLM observability platform for tracing, evaluation, and prompt management.
Weaviate
⭐ 16.0kOpen-source vector database with built-in vectorization and hybrid search.
Ragas
⭐ 13.3kFramework for evaluating RAG pipeline quality with research-backed metrics.
Text Generation Inference (TGI)
⭐ 10.8kHugging Face's production-ready inference server for LLMs.
LangSmith
⭐ 843LangChain's platform for debugging, testing, evaluating, and monitoring LLM applications.
Pinecone
⭐ 429Fully managed vector database designed for high-performance similarity search at scale.
All Projects
- ChromaDB — Simple open-source vector database for AI applications.
- LangFuse — Open-source LLM observability platform for tracing, evaluation, and prompt management.
- LangSmith — LangChain's platform for debugging, testing, evaluating, and monitoring LLM applications.
- Ollama — The easiest way to run open-source LLMs locally.
- Pinecone — Fully managed vector database designed for high-performance similarity search at scale.
- Qdrant — High-performance vector database written in Rust with advanced filtering capabilities.
- Ragas — Framework for evaluating RAG pipeline quality with research-backed metrics.
- Text Generation Inference (TGI) — Hugging Face's production-ready inference server for LLMs.
- vLLM — High-throughput LLM inference engine with PagedAttention for efficient GPU memory management.
- Weaviate — Open-source vector database with built-in vectorization and hybrid search.