llama-cpp: fix postPatch phase
This commit is contained in:
parent
e98a23934a
commit
72d950007b
@ -18,7 +18,13 @@
|
|||||||
port = service_configs.ports.llama_cpp;
|
port = service_configs.ports.llama_cpp;
|
||||||
host = "0.0.0.0";
|
host = "0.0.0.0";
|
||||||
# vulkan broken: https://github.com/ggml-org/llama.cpp/issues/13801
|
# vulkan broken: https://github.com/ggml-org/llama.cpp/issues/13801
|
||||||
package = (lib.optimizePackage inputs.llamacpp.packages.${pkgs.system}.default);
|
package = (
|
||||||
|
lib.optimizePackage (
|
||||||
|
inputs.llamacpp.packages.${pkgs.system}.default.overrideAttrs (old: {
|
||||||
|
postPatch = "";
|
||||||
|
})
|
||||||
|
)
|
||||||
|
);
|
||||||
extraFlags = [
|
extraFlags = [
|
||||||
# "-ngl"
|
# "-ngl"
|
||||||
# "9999"
|
# "9999"
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user