llama-cpp: disable flash attn
This commit is contained in:
parent
a292c2fc75
commit
d0da2591a3
@ -20,7 +20,6 @@
|
|||||||
host = "0.0.0.0";
|
host = "0.0.0.0";
|
||||||
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
|
package = (optimizePackage inputs.llamacpp.packages.${pkgs.system}.vulkan);
|
||||||
extraFlags = [
|
extraFlags = [
|
||||||
"--flash-attn"
|
|
||||||
"-ngl"
|
"-ngl"
|
||||||
"9999"
|
"9999"
|
||||||
];
|
];
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user