llama-cpp: disable flash attn
This commit is contained in:
parent
a292c2fc75
commit
d0da2591a3
@ -20,7 +20,6 @@
|
||||
host = "0.0.0.0";
|
||||
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
|
||||
extraFlags = [
|
||||
"--flash-attn"
|
||||
"-ngl"
|
||||
"9999"
|
||||
];
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user