Skip to main content
Run any Skill in Manus
with one click
$pwd:

vllm-docs

// Use when working with vLLM inference engine: OpenAI-compatible serving, model deployment, quantization (AWQ, GPTQ, FP8, GGUF, INT4/INT8), speculative decoding, LoRA adapters, structured outputs, tool calling, multimodal inputs, distributed serving (tensor/pipeline/expert/context parallel), Docker/Kubernetes deployment, engine configuration, memory optimization, PagedAttention, offline inference, CLI usage, or troubleshooting vLLM issues.

$ git log --oneline --stat
stars:3
forks:0
updated:May 6, 2026 at 11:08
File Explorer
100 files
SKILL.md
readonly