vllm-deployment

Deploy vLLM for high-performance LLM inference. Covers Docker CPU/GPU deployments and cloud VM provisioning with OpenAI-compatible API endpoints.

$ 설치

git clone https://github.com/stakpak/community-paks /tmp/community-paks && cp -r /tmp/community-paks/vllm-deployment ~/.claude/skills/community-paks

// tip: Run this command in your terminal to install the skill