2.7 KiB
2.7 KiB
name, description
| name | description |
|---|---|
| update-vllm-3090-configs | Update only the qwen3.6-27b vLLM 3090 llama-swap configs from club-3090 refs; compare diffs, present a plan, and require approval before editing. |
Update vLLM 3090 Configs
Scope
Use only for Qwen3.6 27B vLLM 3090 configs in modules/nixos/services/llama-swap/.
Do not use this skill for other models, other Qwen sizes, non-vLLM configs, or package bumps.
Local files:
modules/nixos/services/llama-swap/config.nixmodules/nixos/services/llama-swap/setup-qwen36-vllm.sh
Local config keys:
vllm-qwen3.6-27b-tools-textvllm-qwen3.6-27b-long-textvllm-qwen3.6-27b-long-vision
Upstream References
Compare against club-3090 master:
models/qwen3.6-27b/vllm/compose/single/tools-text.ymlmodels/qwen3.6-27b/vllm/compose/single/long-text.ymlmodels/qwen3.6-27b/vllm/compose/single/long-vision.ymlscripts/setup.shfor the currentGENESIS_PIN="${GENESIS_PIN:-...}"
Use raw URLs or a temp clone under _scratch/club-3090. Prefer a temp clone when checking broad changes:
mkdir -p _scratch
git clone https://github.com/noonghunna/club-3090 _scratch/club-3090 2>/dev/null || git -C _scratch/club-3090 pull --ff-only
Required Workflow
- Fetch/update upstream refs under
_scratch/club-3090or fetch the raw files. - Compare upstream compose files to the three local llama-swap entries. Translate docker-compose semantics into the existing
docker run/llama-swap format. - Compare upstream
scripts/setup.shGenesis pin to localGENESIS_PINinsetup-qwen36-vllm.sh. - Check upstream compose volumes/entrypoint for sidecar patches. If patches are added, removed, renamed, or invoked differently, update both:
- runtime mounts and
python3 /patches/...calls inconfig.nix - download/install logic and summary in
setup-qwen36-vllm.sh
- runtime mounts and
- Ignore these diffs unless the user explicitly asks otherwise:
shm_size/ shm-related compose settings- local timing patch
patch_timings_07351e088.pyand its mount/invocation - model served-name differences caused by llama-swap
${MODEL_ID} HUGGING_FACE_HUB_TOKEN; keep local CUDA device/env choices- upstream relative paths vs local
/mnt/ssd/vLLM/...paths - docker-compose format vs local llama-swap/Nix format
- Before editing, present:
- upstream files/commit checked
- meaningful diffs found
- ignored diffs
- exact planned local changes Then wait for explicit user approval.
- After approval, edit minimally and validate:
bash -n modules/nixos/services/llama-swap/setup-qwen36-vllm.shnix-instantiate --parse modules/nixos/services/llama-swap/config.nix
- Summarize changed files and any remaining upstream differences.