Popular Tags
No tags found in this context

flux/cpp

favicongithub.com
url image

Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs. Join our discord: https://discord.gg/5xXzkMu8Zk - lemonade-sdk/lemonade

llmopen-source-coding-agentaiamdgpu
12 hours
favicongithub.com
url image

Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs. - FastFlowLM/FastFlowLM

llmollamaamdllamanpu
4 days
favicongithub.com
url image

Lucebox optimization hub: hand-tuned LLM inference, built for specific consumer hardware. - Luce-Org/lucebox-hub

llmopen-source-coding-agentkernelnvidia-cudam5
10 days
favicongithub.com
url image

MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI. - alibaba/MNN

llmopen-source-coding-agentmachine-learningarmdeep-learning
27 days
favicongithub.com
url image

A high-performance inference engine for LLMs, optimized for diverse AI accelerators. - jd-opensource/xllm

llmopen-source-coding-agentinferenceglminference-engine
27 days
favicongithub.com
url image

LLM inference in C/C++. Contribute to ggml-org/llama.cpp development by creating an account on GitHub.

llmopen-source-coding-agentggmlcpp
27 days