llm --help

Let's help help help devs.

Target: <200ms. Several important LLM CLI tools take multiple seconds. PRs welcome ❤️

library cold warm (10 runs) version measured on
vllm --help15610ms7473ms0.14.1+cpu2026-01-28T21:50Z
sglang --help15662ms6825msv0.5.92026-02-25T11:16Z
VLMEvalKit --help13936ms5076msv0.22026-02-25T11:21Z
tensorrt-llm --help7467ms2600ms1.0.02026-02-25T11:14Z
datasets --help3074ms907ms4.5.02026-02-25T11:02Z
llm --help1205ms569ms0.282026-02-25T11:03Z
openai --help1053ms504ms2.24.02026-02-25T11:03Z
hf --help990ms331ms1.4.12026-02-25T11:02Z
langchain-cli --help845ms259ms0.0.372026-02-25T11:03Z
lm-eval --help800ms228ms0.4.112026-02-25T11:04Z
ollama --help15ms15ms0.17.02026-02-25T11:02Z
llama.cpp --help19ms13msb81492026-02-25T11:04Z
transformers --help0ms0ms5.2.02026-02-25T11:04Z