From dce002cc243588ba5dafbd94909cb9e77fad25de Mon Sep 17 00:00:00 2001 From: Evan Reichard Date: Sun, 28 Dec 2025 09:35:48 -0500 Subject: [PATCH] chore: update llamacpp --- packages/llama-cpp/default.nix | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/packages/llama-cpp/default.nix b/packages/llama-cpp/default.nix index f3d3e8b..ac4c438 100644 --- a/packages/llama-cpp/default.nix +++ b/packages/llama-cpp/default.nix @@ -7,12 +7,12 @@ vulkanSupport = true; }).overrideAttrs (oldAttrs: rec { - version = "7501"; + version = "7562"; src = pkgs.fetchFromGitHub { owner = "ggml-org"; repo = "llama.cpp"; tag = "b${version}"; - hash = "sha256-+bHlQvRUKSoDcULOWqxXWHuYupDmJQuQPg8A7iwsmMY="; + hash = "sha256-yuTPj41B3YitRPrD6yV25ilrIxVKebPGSqdJMpVhUDg="; leaveDotGit = true; postFetch = '' git -C "$out" rev-parse --short HEAD > $out/COMMIT @@ -22,7 +22,6 @@ # Auto CPU Optimizations cmakeFlags = (oldAttrs.cmakeFlags or [ ]) ++ [ - "-DGGML_NATIVE=ON" "-DGGML_CUDA_ENABLE_UNIFIED_MEMORY=1" "-DCMAKE_CUDA_ARCHITECTURES=61;86" # GTX 1070 / GTX 1080ti / RTX 3090 ];