vllm-deployment
Deploy vLLM for high-performance LLM inference. Covers Docker CPU/GPU deployments and cloud VM provisioning with OpenAI-compatible API endpoints.
$ 설치
git clone https://github.com/stakpak/community-paks /tmp/community-paks && cp -r /tmp/community-paks/vllm-deployment ~/.claude/skills/community-paks// tip: Run this command in your terminal to install the skill
