From 4dc577fdcb34f8d8320e167ff983530f2338ca9f Mon Sep 17 00:00:00 2001 From: Simon Gardling Date: Wed, 28 May 2025 21:09:45 -0700 Subject: [PATCH] llama-cpp: disable gpu --- services/llama-cpp.nix | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/services/llama-cpp.nix b/services/llama-cpp.nix index 557faee..cb8fdbc 100644 --- a/services/llama-cpp.nix +++ b/services/llama-cpp.nix @@ -21,8 +21,8 @@ # vulkan broken: https://github.com/ggml-org/llama.cpp/issues/13801 package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.default); extraFlags = [ - "-ngl" - "9999" + # "-ngl" + # "9999" ]; };