From d0da2591a3ae5c943e1454f541a3bb7b29af95ae Mon Sep 17 00:00:00 2001 From: Simon Gardling Date: Wed, 28 May 2025 21:00:08 -0700 Subject: [PATCH] llama-cpp: disable flash attn --- services/llama-cpp.nix | 1 - 1 file changed, 1 deletion(-) diff --git a/services/llama-cpp.nix b/services/llama-cpp.nix index e29e4e8..c49b32c 100644 --- a/services/llama-cpp.nix +++ b/services/llama-cpp.nix @@ -20,7 +20,6 @@ host = "0.0.0.0"; package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan); extraFlags = [ - "--flash-attn" "-ngl" "9999" ];