Community curated code

Ollama enables developers to utilize large language models for efficient AI code creation and improved application security.

FastFlowLM enables efficient execution of large language models on AMD Ryzen AI NPUs, optimizing performance without GPU dependency.
Shimmy is a Rust-based inference server providing local, OpenAI-compatible endpoints for machine learning models.
SGLang is an open-source framework for efficient serving of large language and multimodal models, ensuring low-latency and high-throughput performance.
vLLM is an efficient engine for LLM inference and serving, designed for high throughput and memory management.

Oumi is an open-source platform for training and deploying LLMs and VLMs, providing tools for evaluation and data synthesis.

Unsloth is a web UI for training and running AI models locally, enhancing efficiency and performance.

LocalAI is an open-source AI engine for running models locally without GPU requirements, ensuring privacy and flexibility.

LangChain4j simplifies LLM integration in Java applications with a unified API and a comprehensive toolbox for developers.