480a8106cf
Use redb instead of sled for the default filesystem implementation of PathInfoService and DirectoryService. In the future we'll also drop sled support completely. Change-Id: I513ff0c2ff953d59714aa50b9aa1301b02f53d40 Reviewed-on: https://cl.tvl.fyi/c/depot/+/12085 Autosubmit: Ilan Joselevich <personal@ilanjoselevich.com> Reviewed-by: flokli <flokli@flokli.de> Tested-by: BuildkiteCI
231 lines
9.7 KiB
Nix
231 lines
9.7 KiB
Nix
{ depot, pkgs, lib, ... }:
|
|
|
|
let
|
|
# Seed a tvix-store with the tvix docs, then start a VM, ask it to list all
|
|
# files in /nix/store, and ensure the store path is present, which acts as a
|
|
# nice smoketest.
|
|
mkBootTest =
|
|
{ blobServiceAddr ? "memory://"
|
|
, directoryServiceAddr ? "memory://"
|
|
, pathInfoServiceAddr ? "memory://"
|
|
|
|
|
|
# The path to import.
|
|
, path
|
|
|
|
# Whether the path should be imported as a closure.
|
|
# If false, importPathName must be specified.
|
|
, isClosure ? false
|
|
# Whether to use nar-bridge to upload, rather than tvix-store copy.
|
|
# using nar-bridge currently is "slower", as the `pkgs.mkBinaryCache` build
|
|
# takes quite some time.
|
|
, useNarBridge ? false
|
|
|
|
, importPathName ? null
|
|
|
|
# Commands to run before starting the tvix-daemon. Useful to provide
|
|
# auxillary mock services.
|
|
, preStart ? ""
|
|
|
|
# The cmdline to pass to the VM.
|
|
# Defaults to tvix.find, which lists all files in the store.
|
|
, vmCmdline ? "tvix.find"
|
|
# The string we expect to find in the VM output.
|
|
# Defaults the value of `path` (the store path we upload).
|
|
, assertVMOutput ? path
|
|
}:
|
|
|
|
assert isClosure -> importPathName == null;
|
|
assert (!isClosure) -> importPathName != null;
|
|
|
|
pkgs.stdenv.mkDerivation ({
|
|
name = "run-vm";
|
|
|
|
nativeBuildInputs = [
|
|
depot.tvix.store
|
|
depot.tvix.boot.runVM
|
|
] ++ lib.optionals (isClosure && useNarBridge) [
|
|
depot.tvix.nar-bridge
|
|
pkgs.curl
|
|
pkgs.parallel
|
|
pkgs.xz.bin
|
|
];
|
|
buildCommand = ''
|
|
touch $out
|
|
# Ensure we can construct http clients.
|
|
export SSL_CERT_FILE=/dev/null
|
|
|
|
${preStart}
|
|
|
|
# Start the tvix daemon, listening on a unix socket.
|
|
BLOB_SERVICE_ADDR=${lib.escapeShellArg blobServiceAddr} \
|
|
DIRECTORY_SERVICE_ADDR=${lib.escapeShellArg directoryServiceAddr} \
|
|
PATH_INFO_SERVICE_ADDR=${lib.escapeShellArg pathInfoServiceAddr} \
|
|
tvix-store \
|
|
--otlp=false \
|
|
daemon -l $PWD/tvix-store.sock &
|
|
|
|
# Wait for the service to report healthy.
|
|
timeout 22 sh -c "until ${pkgs.ip2unix}/bin/ip2unix -r out,path=$PWD/tvix-store.sock ${pkgs.grpc-health-check}/bin/grpc-health-check --address 127.0.0.1 --port 8080; do sleep 1; done"
|
|
|
|
# Export env vars so that subsequent tvix-store commands will talk to
|
|
# our tvix-store daemon over the unix socket.
|
|
export BLOB_SERVICE_ADDR=grpc+unix://$PWD/tvix-store.sock
|
|
export DIRECTORY_SERVICE_ADDR=grpc+unix://$PWD/tvix-store.sock
|
|
export PATH_INFO_SERVICE_ADDR=grpc+unix://$PWD/tvix-store.sock
|
|
'' + lib.optionalString (!isClosure) ''
|
|
echo "Importing ${path} into tvix-store with name ${importPathName}…"
|
|
cp -R ${path} ${importPathName}
|
|
outpath=$(tvix-store import ${importPathName})
|
|
|
|
echo "imported to $outpath"
|
|
'' + lib.optionalString (isClosure && !useNarBridge) ''
|
|
echo "Copying closure ${path}…"
|
|
# This picks up the `closure` key in `$NIX_ATTRS_JSON_FILE` automatically.
|
|
tvix-store --otlp=false copy
|
|
'' + lib.optionalString (isClosure && useNarBridge) ''
|
|
echo "Starting nar-bridge…"
|
|
nar-bridge \
|
|
--otlp=false \
|
|
-l $PWD/nar-bridge.sock &
|
|
|
|
# Wait for nar-bridge to report healthy.
|
|
timeout 22 sh -c "until ${pkgs.curl}/bin/curl -s --unix-socket $PWD/nar-bridge.sock http:///nix-binary-cache; do sleep 1; done"
|
|
|
|
# Upload. We can't use nix copy --to http://…, as it wants access to the nix db.
|
|
# However, we can use mkBinaryCache to assemble .narinfo and .nar.xz to upload,
|
|
# and then drive a HTTP client ourselves.
|
|
to_upload=${pkgs.mkBinaryCache { rootPaths = [path];}}
|
|
|
|
# Upload all NAR files (with some parallelism).
|
|
# As mkBinaryCache produces them xz-compressed, unpack them on the fly.
|
|
# nar-bridge doesn't care about the path we upload *to*, but a
|
|
# subsequent .narinfo upload need to refer to its contents (by narhash).
|
|
echo -e "Uploading NARs… "
|
|
ls -d $to_upload/nar/*.nar.xz | parallel 'xz -d < {} | curl -s -T - --unix-socket $PWD/nar-bridge.sock http://localhost:9000/nar/$(basename {} | cut -d "." -f 1).nar'
|
|
echo "Done."
|
|
|
|
# Upload all NARInfo files.
|
|
# FUTUREWORK: This doesn't upload them in order, and currently relies
|
|
# on PathInfoService not doing any checking.
|
|
# In the future, we might want to make this behaviour configurable,
|
|
# and disable checking here, to keep the logic simple.
|
|
ls -d $to_upload/*.narinfo | parallel 'curl -s -T - --unix-socket $PWD/nar-bridge.sock http://localhost:9000/$(basename {}) < {}'
|
|
'' + ''
|
|
# Invoke a VM using tvix as the backing store, ensure the outpath appears in its listing.
|
|
echo "Starting VM…"
|
|
|
|
CH_CMDLINE="${vmCmdline}" run-tvix-vm 2>&1 | tee output.txt
|
|
grep "${assertVMOutput}" output.txt
|
|
'';
|
|
requiredSystemFeatures = [ "kvm" ];
|
|
} // lib.optionalAttrs (isClosure && !useNarBridge) {
|
|
__structuredAttrs = true;
|
|
exportReferencesGraph.closure = [ path ];
|
|
});
|
|
|
|
systemFor = sys: (depot.ops.nixos.nixosFor sys).system;
|
|
|
|
testSystem = systemFor ({ modulesPath, pkgs, ... }: {
|
|
# Set some options necessary to evaluate.
|
|
boot.loader.systemd-boot.enable = true;
|
|
# TODO: figure out how to disable this without causing eval to fail
|
|
fileSystems."/" = {
|
|
device = "/dev/root";
|
|
fsType = "tmpfs";
|
|
};
|
|
|
|
services.getty.helpLine = "Onwards and upwards.";
|
|
systemd.services.do-shutdown = {
|
|
after = [ "getty.target" ];
|
|
description = "Shut down again";
|
|
wantedBy = [ "multi-user.target" ];
|
|
serviceConfig.Type = "oneshot";
|
|
script = "/run/current-system/sw/bin/systemctl poweroff --when=+10s";
|
|
};
|
|
|
|
# Don't warn about stateVersion.
|
|
system.stateVersion = "24.05";
|
|
});
|
|
|
|
in
|
|
depot.nix.readTree.drvTargets
|
|
{
|
|
docs-memory = (mkBootTest {
|
|
path = ../../docs;
|
|
importPathName = "docs";
|
|
});
|
|
docs-persistent = (mkBootTest {
|
|
blobServiceAddr = "objectstore+file:///build/blobs";
|
|
directoryServiceAddr = "redb:///build/directories.redb";
|
|
pathInfoServiceAddr = "redb:///build/pathinfo.redb";
|
|
path = ../../docs;
|
|
importPathName = "docs";
|
|
});
|
|
|
|
closure-tvix = (mkBootTest {
|
|
blobServiceAddr = "objectstore+file:///build/blobs";
|
|
path = depot.tvix.store;
|
|
isClosure = true;
|
|
});
|
|
|
|
closure-nixos = (mkBootTest {
|
|
blobServiceAddr = "objectstore+file:///build/blobs";
|
|
pathInfoServiceAddr = "redb:///build/pathinfo.redb";
|
|
directoryServiceAddr = "redb:///build/directories.redb";
|
|
path = testSystem;
|
|
isClosure = true;
|
|
vmCmdline = "init=${testSystem}/init panic=-1"; # reboot immediately on panic
|
|
assertVMOutput = "Onwards and upwards.";
|
|
});
|
|
|
|
closure-nixos-bigtable = (mkBootTest {
|
|
blobServiceAddr = "objectstore+file:///build/blobs";
|
|
directoryServiceAddr = "bigtable://instance-1?project_id=project-1&table_name=directories&family_name=cf1";
|
|
pathInfoServiceAddr = "bigtable://instance-1?project_id=project-1&table_name=pathinfos&family_name=cf1";
|
|
path = testSystem;
|
|
useNarBridge = true;
|
|
preStart = ''
|
|
${pkgs.cbtemulator}/bin/cbtemulator -address $PWD/cbtemulator.sock &
|
|
timeout 22 sh -c 'until [ -e $PWD/cbtemulator.sock ]; do sleep 1; done'
|
|
|
|
export BIGTABLE_EMULATOR_HOST=unix://$PWD/cbtemulator.sock
|
|
${pkgs.google-cloud-bigtable-tool}/bin/cbt -instance instance-1 -project project-1 createtable directories
|
|
${pkgs.google-cloud-bigtable-tool}/bin/cbt -instance instance-1 -project project-1 createfamily directories cf1
|
|
${pkgs.google-cloud-bigtable-tool}/bin/cbt -instance instance-1 -project project-1 createtable pathinfos
|
|
${pkgs.google-cloud-bigtable-tool}/bin/cbt -instance instance-1 -project project-1 createfamily pathinfos cf1
|
|
'';
|
|
isClosure = true;
|
|
vmCmdline = "init=${testSystem}/init panic=-1"; # reboot immediately on panic
|
|
assertVMOutput = "Onwards and upwards.";
|
|
});
|
|
|
|
closure-nixos-s3 = (mkBootTest {
|
|
blobServiceAddr = "objectstore+s3://mybucket/blobs?aws_access_key_id=myaccesskey&aws_secret_access_key=supersecret&aws_endpoint_url=http%3A%2F%2Flocalhost%3A9000&aws_allow_http=1";
|
|
# we cannot use s3 here yet without any caching layer, as we don't allow "deeper" access to directories (non-root nodes)
|
|
# directoryServiceAddr = "objectstore+s3://mybucket/directories?aws_access_key_id=myaccesskey&aws_secret_access_key=supersecret&endpoint=http%3A%2F%2Flocalhost%3A9000&aws_allow_http=1";
|
|
directoryServiceAddr = "memory://";
|
|
pathInfoServiceAddr = "memory://";
|
|
path = testSystem;
|
|
useNarBridge = true;
|
|
preStart = ''
|
|
MINIO_ACCESS_KEY=myaccesskey MINIO_SECRET_KEY=supersecret MINIO_ADDRESS=127.0.0.1:9000 ${pkgs.minio}/bin/minio server $(mktemp -d) &
|
|
timeout 22 sh -c 'until ${pkgs.netcat}/bin/nc -z $0 $1; do sleep 1; done' localhost 9000
|
|
mc_config_dir=$(mktemp -d)
|
|
${pkgs.minio-client}/bin/mc --config-dir $mc_config_dir alias set 'myminio' 'http://127.0.0.1:9000' 'myaccesskey' 'supersecret'
|
|
${pkgs.minio-client}/bin/mc --config-dir $mc_config_dir mb myminio/mybucket
|
|
'';
|
|
isClosure = true;
|
|
vmCmdline = "init=${testSystem}/init panic=-1"; # reboot immediately on panic
|
|
assertVMOutput = "Onwards and upwards.";
|
|
});
|
|
|
|
closure-nixos-nar-bridge = (mkBootTest {
|
|
blobServiceAddr = "objectstore+file:///build/blobs";
|
|
path = testSystem;
|
|
useNarBridge = true;
|
|
isClosure = true;
|
|
vmCmdline = "init=${testSystem}/init panic=-1"; # reboot immediately on panic
|
|
assertVMOutput = "Onwards and upwards.";
|
|
});
|
|
}
|