llama-cpp: disable flash attn
This commit is contained in:
@@ -20,7 +20,6 @@
|
||||
host = "0.0.0.0";
|
||||
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
|
||||
extraFlags = [
|
||||
"--flash-attn"
|
||||
"-ngl"
|
||||
"9999"
|
||||
];
|
||||
|
||||
Reference in New Issue
Block a user