chore: update llama-cpp to b7898 and opencode to v1.1.48

- Update llama-cpp from b7867 to b7898
- Update opencode from v1.1.12 to v1.1.48 with improved build process:
  - Replace custom bundle script with official script/build.ts
  - Add shell completion support
  - Add version check testing
  - Simplify node_modules handling
- Update llama-swap service config with new llama.cpp options
- Clarify opencode agent testing workflow in developer and reviewer configs
This commit is contained in:
2026-02-03 20:33:14 -05:00
parent 15259b5ec3
commit 7080727dce
8 changed files with 174 additions and 163 deletions

View File

@@ -19,8 +19,8 @@ in
--port ''${PORT} \
-m /mnt/ssd/Models/GLM/GLM-4.7-Flash-UD-Q4_K_XL.gguf \
-c ''${ctx} \
-ctk q8_0 \
--jinja \
--threads -1 \
--temp 0.7 \
--top-p 1.0 \
--min-p 0.01 \
@@ -34,6 +34,11 @@ in
env = [ "GGML_CUDA_ENABLE_UNIFIED_MEMORY=1" ];
};
# --spec-type ngram-mod \
# --spec-ngram-size-n 24 \
# --draft-min 48 \
# --draft-max 64 \
# https://huggingface.co/unsloth/Devstral-Small-2-24B-Instruct-2512-GGUF/tree/main
"devstral-small-2-instruct" = {
name = "Devstral Small 2 (24B) - Instruct";