Side-by-side comparison of stars, features, and trends
LiteLLM is an open-source AI gateway that provides a unified interface for calling over 100 different LLM providers using the standard OpenAI format. It can be utilized as a Python SDK for direct integration or deployed as a proxy server to manage enterprise-grade features like load balancing and spend tracking. By abstracting provider-specific complexities, it allows developers to switch between models seamlessly without rewriting their existing code.
DeepGEMM is a unified CUDA library providing high-performance tensor core kernels specifically optimized for modern large language models. It features a lightweight Just-In-Time compilation module that eliminates the need for complex CUDA installation builds. The library supports advanced operations including FP8/FP4 GEMMs, fused Mega MoE, and MQA scoring kernels for efficient GPU computation.