Popular Tags
No tags found in this context
Community curated code
Shimmy is a Rust-based inference server providing local, OpenAI-compatible endpoints for machine learning models.

xLLM is an efficient inference engine for large language models, optimized for AI accelerators, enabling cost-effective enterprise deployment.