From cd668264a5d8295bc514eedd4da704ade2e420c9 Mon Sep 17 00:00:00 2001 From: ant Date: Sat, 24 Jan 2026 15:15:44 +0100 Subject: [PATCH] disable llama things --- os/basado/configuration.nix | 35 ++++++++++++++++++++++++++++------- 1 file changed, 28 insertions(+), 7 deletions(-) diff --git a/os/basado/configuration.nix b/os/basado/configuration.nix index 9b34986..0078dcf 100644 --- a/os/basado/configuration.nix +++ b/os/basado/configuration.nix @@ -87,13 +87,34 @@ in virtualisation.libvirtd.enable = true; programs.virt-manager.enable = true; - services.ollama = { - enable = true; - acceleration = "cuda"; - }; - services.open-webui = { - enable = true; - }; + # services.llama-cpp = { + # package = pkgs-unstable.llama-cpp-vulkan; + # enable = true; + # port = 8182; + # model = "/home/ant/models/Qwen3-4B-Instruct-2507-UD-Q8_K_XL.gguf"; + # extraFlags = [ + # "-fa" "on" + # "--jinja" + # "--cache-reuse" "256" + # "--reasoning-format" "auto" + # "--ctx-size" "16384" + # "--n-gpu-layers" "999" + # "--cache-type-k" "q4_0" + # "--cache-type-v" "q4_0" + # "--n-cpu-moe" "25" + # ]; + # }; + # systemd.services.llama-cpp = { + # environment.XDG_CACHE_HOME = "/var/cache/llama.cpp"; + # serviceConfig.CacheDirectory = "llama.cpp"; + # }; + # services.ollama = { + # enable = true; + # acceleration = "cuda"; + # }; + # services.open-webui = { + # enable = true; + # }; services.immich.enable = true; # This value determines the NixOS release from which the default