llama-cpp: disable flash attn

This commit is contained in:
Simon Gardling 2025-05-28 21:00:08 -07:00
parent a292c2fc75
commit d0da2591a3
Signed by: titaniumtown
GPG Key ID: 9AB28AC10ECE533D

View File

@ -20,7 +20,6 @@
host = "0.0.0.0";
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
extraFlags = [
"--flash-attn"
"-ngl"
"9999"
];