llama-cpp: fix postPatch phase
This commit is contained in:
parent
e98a23934a
commit
72d950007b
@ -18,7 +18,13 @@
|
||||
port = service_configs.ports.llama_cpp;
|
||||
host = "0.0.0.0";
|
||||
# vulkan broken: https://github.com/ggml-org/llama.cpp/issues/13801
|
||||
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.default);
|
||||
package = (
|
||||
lib.optimizePackage (
|
||||
inputs.llamacpp.packages.${pkgs.system}.default.overrideAttrs (old: {
|
||||
postPatch = "";
|
||||
})
|
||||
)
|
||||
);
|
||||
extraFlags = [
|
||||
# "-ngl"
|
||||
# "9999"
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user