forked from DGNum/infrastructure
feat(krz01): add a NGINX in front of ollama protected by password
This way, you can do direct requests to ollama from other places. Signed-off-by: Ryan Lahfa <ryan@dgnum.eu>
This commit is contained in:
parent
7d70beb1f0
commit
d76e655174
2 changed files with 33 additions and 8 deletions
|
@ -12,11 +12,11 @@ lib.extra.mkConfig {
|
||||||
|
|
||||||
enabledServices = [
|
enabledServices = [
|
||||||
# INFO: This list needs to stay sorted alphabetically
|
# INFO: This list needs to stay sorted alphabetically
|
||||||
|
# Machine learning API machine
|
||||||
|
"microvm-ml01"
|
||||||
"microvm-router01"
|
"microvm-router01"
|
||||||
"nvidia-tesla-k80"
|
"nvidia-tesla-k80"
|
||||||
"proxmox"
|
"proxmox"
|
||||||
# Machine learning API machine
|
|
||||||
"microvm-ml01"
|
|
||||||
];
|
];
|
||||||
|
|
||||||
extraConfig = {
|
extraConfig = {
|
||||||
|
@ -57,12 +57,29 @@ lib.extra.mkConfig {
|
||||||
];
|
];
|
||||||
}))
|
}))
|
||||||
];
|
];
|
||||||
services.ollama = {
|
|
||||||
enable = true;
|
services = {
|
||||||
package = pkgs.callPackage ./ollama.nix {
|
nginx = {
|
||||||
cudaPackages = pkgs.cudaPackages_11;
|
enable = true;
|
||||||
# We need to thread our nvidia x11 driver for CUDA.
|
recommendedProxySettings = true;
|
||||||
extraLibraries = [ config.hardware.nvidia.package ];
|
virtualHosts."ollama01.beta.dgnum.eu" = {
|
||||||
|
enableACME = true;
|
||||||
|
forceSSL = true;
|
||||||
|
locations."/" = {
|
||||||
|
proxyPass = "http://${config.services.ollama.host}:${toString config.services.ollama.port}";
|
||||||
|
basicAuthFile = pkgs.writeText "ollama-htpasswd" ''
|
||||||
|
raito:$y$j9T$UDEHpLtM52hRGK0I4qT6M0$N75AhENLqgtJnTGaPzq51imhjZvuPr.ow81Co1ZTcX2
|
||||||
|
'';
|
||||||
|
};
|
||||||
|
};
|
||||||
|
};
|
||||||
|
ollama = {
|
||||||
|
enable = true;
|
||||||
|
package = pkgs.callPackage ./ollama.nix {
|
||||||
|
cudaPackages = pkgs.cudaPackages_11;
|
||||||
|
# We need to thread our nvidia x11 driver for CUDA.
|
||||||
|
extraLibraries = [ config.hardware.nvidia.package ];
|
||||||
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|
|
@ -127,6 +127,14 @@ let
|
||||||
"cas-eleves"
|
"cas-eleves"
|
||||||
"vote"
|
"vote"
|
||||||
];
|
];
|
||||||
|
|
||||||
|
krz01.dual = [
|
||||||
|
# Beta-grade machine learning API servers
|
||||||
|
"ollama01.beta"
|
||||||
|
"openui.beta"
|
||||||
|
"whisper.beta"
|
||||||
|
"stable-diffusion.beta"
|
||||||
|
];
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
);
|
);
|
||||||
|
|
Loading…
Reference in a new issue