
whichllm helps you find the best local LLM for your hardware, optimizing AI inference with real-time benchmarks.
whichllm is a command-line tool designed to identify the best local large language model (LLM) that operates efficiently on your hardware. It automatically detects your GPU, CPU, and RAM specifications, and ranks models from HuggingFace based on real performance benchmarks rather than just parameter counts.
Key features:
This tool is particularly useful for developers and researchers looking to optimize their AI inference tasks by selecting the most suitable model for their specific hardware configuration.