refactor(tvix/nar-bridge): let callbaks return calculated digests

This aligns behaviour more with how it should be - it's the
responsibility of the callback functions to return digests of the things
they consume(d). It allows further cleaning up the hasher struct.

Change-Id: I9cbfc87e6abd4ff17fadf39eb6563ec3cb7fcc6f
Reviewed-on: https://cl.tvl.fyi/c/depot/+/9528
Autosubmit: flokli <flokli@flokli.de>
Tested-by: BuildkiteCI
Reviewed-by: Connor Brewster <cbrewster@hey.com>
This commit is contained in:
Florian Klink 2023-10-03 13:59:13 +03:00 committed by flokli
parent b1ff1267be
commit f92b0ef933
6 changed files with 109 additions and 100 deletions

View file

@ -30,9 +30,9 @@ func Import(
// The reader the data is read from // The reader the data is read from
r io.Reader, r io.Reader,
// callback function called with each regular file content // callback function called with each regular file content
blobCb func(fileReader io.Reader) error, blobCb func(fileReader io.Reader) ([]byte, error),
// callback function called with each finalized directory node // callback function called with each finalized directory node
directoryCb func(directory *castorev1pb.Directory) error, directoryCb func(directory *castorev1pb.Directory) ([]byte, error),
) (*storev1pb.PathInfo, error) { ) (*storev1pb.PathInfo, error) {
// wrap the passed reader in a reader that records the number of bytes read and // wrap the passed reader in a reader that records the number of bytes read and
// their sha256 sum. // their sha256 sum.
@ -65,24 +65,21 @@ func Import(
toPop := stack[len(stack)-1] toPop := stack[len(stack)-1]
stack = stack[:len(stack)-1] stack = stack[:len(stack)-1]
// if there's still a parent left on the stack, refer to it from there. // call the directoryCb
if len(stack) > 0 { directoryDigest, err := directoryCb(toPop.directory)
dgst, err := toPop.directory.Digest()
if err != nil { if err != nil {
return fmt.Errorf("unable to calculate directory digest: %w", err) return fmt.Errorf("failed calling directoryCb: %w", err)
} }
// if there's still a parent left on the stack, refer to it from there.
if len(stack) > 0 {
topOfStack := stack[len(stack)-1].directory topOfStack := stack[len(stack)-1].directory
topOfStack.Directories = append(topOfStack.Directories, &castorev1pb.DirectoryNode{ topOfStack.Directories = append(topOfStack.Directories, &castorev1pb.DirectoryNode{
Name: []byte(path.Base(toPop.path)), Name: []byte(path.Base(toPop.path)),
Digest: dgst, Digest: directoryDigest,
Size: toPop.directory.Size(), Size: toPop.directory.Size(),
}) })
} }
// call the directoryCb
if err := directoryCb(toPop.directory); err != nil {
return fmt.Errorf("failed calling directoryCb: %w", err)
}
// Keep track that we have encounter at least one directory // Keep track that we have encounter at least one directory
stackDirectory = toPop.directory stackDirectory = toPop.directory
return nil return nil
@ -106,7 +103,7 @@ func Import(
hdr, err := narReader.Next() hdr, err := narReader.Next()
// If this returns an error, it's either EOF (when we're done reading from the NAR), // If this returns an error, it's either EOF (when we're done reading from the NAR),
// or another error // or another error.
if err != nil { if err != nil {
// if this returns no EOF, bail out // if this returns no EOF, bail out
if !errors.Is(err, io.EOF) { if !errors.Is(err, io.EOF) {
@ -206,28 +203,22 @@ func Import(
} }
if hdr.Type == nar.TypeRegular { if hdr.Type == nar.TypeRegular {
// wrap reader with a reader calculating the blake3 hash // wrap reader with a reader calculating the blake3 hash
fileReader := hashers.NewHasher(narReader, blake3.New(32, nil)) blobReader := hashers.NewHasher(narReader, blake3.New(32, nil))
err := blobCb(fileReader) blobDigest, err := blobCb(blobReader)
if err != nil { if err != nil {
return nil, fmt.Errorf("failure from blobCb: %w", err) return nil, fmt.Errorf("failure from blobCb: %w", err)
} }
// drive the file reader to the end, in case the CB function doesn't read // ensure blobCb did read all the way to the end.
// all the way to the end on its own // If it didn't, the blobCb function is wrong and we should bail out.
if fileReader.BytesWritten() != uint32(hdr.Size) { if blobReader.BytesWritten() != uint32(hdr.Size) {
_, err := io.ReadAll(fileReader) panic("not read to end")
if err != nil {
return nil, fmt.Errorf("unable to read until the end of the file content: %w", err)
} }
}
// read the blake3 hash
dgst := fileReader.Sum(nil)
fileNode := &castorev1pb.FileNode{ fileNode := &castorev1pb.FileNode{
Name: []byte(getBasename(hdr.Path)), Name: []byte(getBasename(hdr.Path)),
Digest: dgst, Digest: blobDigest,
Size: uint32(hdr.Size), Size: uint32(hdr.Size),
Executable: hdr.Executable, Executable: hdr.Executable,
} }

View file

@ -1,6 +1,7 @@
package importer_test package importer_test
import ( import (
"bytes"
"context" "context"
"errors" "errors"
"io" "io"
@ -13,6 +14,7 @@ import (
"github.com/google/go-cmp/cmp" "github.com/google/go-cmp/cmp"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
"google.golang.org/protobuf/testing/protocmp" "google.golang.org/protobuf/testing/protocmp"
"lukechampine.com/blake3"
) )
func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) { func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) {
@ -21,7 +23,7 @@ func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) {
} }
} }
func mustDigest(d *castorev1pb.Directory) []byte { func mustDirectoryDigest(d *castorev1pb.Directory) []byte {
dgst, err := d.Digest() dgst, err := d.Digest()
if err != nil { if err != nil {
panic(err) panic(err)
@ -29,6 +31,15 @@ func mustDigest(d *castorev1pb.Directory) []byte {
return dgst return dgst
} }
func mustBlobDigest(r io.Reader) []byte {
hasher := blake3.New(32, nil)
_, err := io.Copy(hasher, r)
if err != nil {
panic(err)
}
return hasher.Sum([]byte{})
}
func TestSymlink(t *testing.T) { func TestSymlink(t *testing.T) {
f, err := os.Open("../../testdata/symlink.nar") f, err := os.Open("../../testdata/symlink.nar")
require.NoError(t, err) require.NoError(t, err)
@ -36,9 +47,9 @@ func TestSymlink(t *testing.T) {
actualPathInfo, err := importer.Import( actualPathInfo, err := importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
panic("no file contents expected!") panic("no file contents expected!")
}, func(directory *castorev1pb.Directory) error { }, func(directory *castorev1pb.Directory) ([]byte, error) {
panic("no directories expected!") panic("no directories expected!")
}, },
) )
@ -74,12 +85,12 @@ func TestRegular(t *testing.T) {
actualPathInfo, err := importer.Import( actualPathInfo, err := importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
contents, err := io.ReadAll(fileReader) contents, err := io.ReadAll(blobReader)
require.NoError(t, err, "reading fileReader should not error") require.NoError(t, err, "reading blobReader should not error")
require.Equal(t, []byte{0x01}, contents, "contents read from fileReader should match expectations") require.Equal(t, []byte{0x01}, contents, "contents read from blobReader should match expectations")
return nil return mustBlobDigest(bytes.NewBuffer(contents)), nil
}, func(directory *castorev1pb.Directory) error { }, func(directory *castorev1pb.Directory) ([]byte, error) {
panic("no directories expected!") panic("no directories expected!")
}, },
) )
@ -129,11 +140,11 @@ func TestEmptyDirectory(t *testing.T) {
actualPathInfo, err := importer.Import( actualPathInfo, err := importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
panic("no file contents expected!") panic("no file contents expected!")
}, func(directory *castorev1pb.Directory) error { }, func(directory *castorev1pb.Directory) ([]byte, error) {
requireProtoEq(t, expectedDirectory, directory) requireProtoEq(t, expectedDirectory, directory)
return nil return mustDirectoryDigest(directory), nil
}, },
) )
require.NoError(t, err) require.NoError(t, err)
@ -143,7 +154,7 @@ func TestEmptyDirectory(t *testing.T) {
Node: &castorev1pb.Node_Directory{ Node: &castorev1pb.Node_Directory{
Directory: &castorev1pb.DirectoryNode{ Directory: &castorev1pb.DirectoryNode{
Name: []byte(""), Name: []byte(""),
Digest: mustDigest(expectedDirectory), Digest: mustDirectoryDigest(expectedDirectory),
Size: expectedDirectory.Size(), Size: expectedDirectory.Size(),
}, },
}, },
@ -415,17 +426,17 @@ func TestFull(t *testing.T) {
Directories: []*castorev1pb.DirectoryNode{ Directories: []*castorev1pb.DirectoryNode{
{ {
Name: []byte("man1"), Name: []byte("man1"),
Digest: mustDigest(expectedDirectories["/share/man/man1"]), Digest: mustDirectoryDigest(expectedDirectories["/share/man/man1"]),
Size: expectedDirectories["/share/man/man1"].Size(), Size: expectedDirectories["/share/man/man1"].Size(),
}, },
{ {
Name: []byte("man5"), Name: []byte("man5"),
Digest: mustDigest(expectedDirectories["/share/man/man5"]), Digest: mustDirectoryDigest(expectedDirectories["/share/man/man5"]),
Size: expectedDirectories["/share/man/man5"].Size(), Size: expectedDirectories["/share/man/man5"].Size(),
}, },
{ {
Name: []byte("man8"), Name: []byte("man8"),
Digest: mustDigest(expectedDirectories["/share/man/man8"]), Digest: mustDirectoryDigest(expectedDirectories["/share/man/man8"]),
Size: expectedDirectories["/share/man/man8"].Size(), Size: expectedDirectories["/share/man/man8"].Size(),
}, },
}, },
@ -438,7 +449,7 @@ func TestFull(t *testing.T) {
Directories: []*castorev1pb.DirectoryNode{ Directories: []*castorev1pb.DirectoryNode{
{ {
Name: []byte("man"), Name: []byte("man"),
Digest: mustDigest(expectedDirectories["/share/man"]), Digest: mustDirectoryDigest(expectedDirectories["/share/man"]),
Size: expectedDirectories["/share/man"].Size(), Size: expectedDirectories["/share/man"].Size(),
}, },
}, },
@ -451,12 +462,12 @@ func TestFull(t *testing.T) {
Directories: []*castorev1pb.DirectoryNode{ Directories: []*castorev1pb.DirectoryNode{
{ {
Name: []byte("bin"), Name: []byte("bin"),
Digest: mustDigest(expectedDirectories["/bin"]), Digest: mustDirectoryDigest(expectedDirectories["/bin"]),
Size: expectedDirectories["/bin"].Size(), Size: expectedDirectories["/bin"].Size(),
}, },
{ {
Name: []byte("share"), Name: []byte("share"),
Digest: mustDigest(expectedDirectories["/share"]), Digest: mustDirectoryDigest(expectedDirectories["/share"]),
Size: expectedDirectories["/share"].Size(), Size: expectedDirectories["/share"].Size(),
}, },
}, },
@ -476,14 +487,12 @@ func TestFull(t *testing.T) {
actualPathInfo, err := importer.Import( actualPathInfo, err := importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
// Don't really bother reading and comparing the contents here, // Don't really bother reading and comparing the contents here,
// We already verify the right digests are produced by comparing the // We already verify the right digests are produced by comparing the
// directoryCb calls, and TestRegular ensures the reader works. // directoryCb calls, and TestRegular ensures the reader works.
// This also covers the case when the client doesn't read from the reader, and that the return mustBlobDigest(blobReader), nil
// importer will take care of reading all the way to the end no matter what. }, func(directory *castorev1pb.Directory) ([]byte, error) {
return nil
}, func(directory *castorev1pb.Directory) error {
// use actualDirectoryOrder to look up the Directory object we expect at this specific invocation. // use actualDirectoryOrder to look up the Directory object we expect at this specific invocation.
currentDirectoryPath := expectedDirectoryPaths[numDirectoriesReceived] currentDirectoryPath := expectedDirectoryPaths[numDirectoriesReceived]
@ -493,7 +502,7 @@ func TestFull(t *testing.T) {
requireProtoEq(t, expectedDirectory, directory) requireProtoEq(t, expectedDirectory, directory)
numDirectoriesReceived += 1 numDirectoriesReceived += 1
return nil return mustDirectoryDigest(directory), nil
}, },
) )
require.NoError(t, err) require.NoError(t, err)
@ -503,7 +512,7 @@ func TestFull(t *testing.T) {
Node: &castorev1pb.Node_Directory{ Node: &castorev1pb.Node_Directory{
Directory: &castorev1pb.DirectoryNode{ Directory: &castorev1pb.DirectoryNode{
Name: []byte(""), Name: []byte(""),
Digest: mustDigest(expectedDirectories["/"]), Digest: mustDirectoryDigest(expectedDirectories["/"]),
Size: expectedDirectories["/"].Size(), Size: expectedDirectories["/"].Size(),
}, },
}, },
@ -524,7 +533,7 @@ func TestFull(t *testing.T) {
// TestCallbackErrors ensures that errors returned from the callback function // TestCallbackErrors ensures that errors returned from the callback function
// bubble up to the importer process, and are not ignored. // bubble up to the importer process, and are not ignored.
func TestCallbackErrors(t *testing.T) { func TestCallbackErrors(t *testing.T) {
t.Run("callback file", func(t *testing.T) { t.Run("callback blob", func(t *testing.T) {
// Pick an example NAR with a regular file. // Pick an example NAR with a regular file.
f, err := os.Open("../../testdata/onebyteregular.nar") f, err := os.Open("../../testdata/onebyteregular.nar")
require.NoError(t, err) require.NoError(t, err)
@ -534,9 +543,9 @@ func TestCallbackErrors(t *testing.T) {
_, err = importer.Import( _, err = importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
return targetErr return nil, targetErr
}, func(directory *castorev1pb.Directory) error { }, func(directory *castorev1pb.Directory) ([]byte, error) {
panic("no directories expected!") panic("no directories expected!")
}, },
) )
@ -552,10 +561,10 @@ func TestCallbackErrors(t *testing.T) {
_, err = importer.Import( _, err = importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
panic("no file contents expected!") panic("no file contents expected!")
}, func(directory *castorev1pb.Directory) error { }, func(directory *castorev1pb.Directory) ([]byte, error) {
return targetErr return nil, targetErr
}, },
) )
require.ErrorIs(t, err, targetErr) require.ErrorIs(t, err, targetErr)
@ -582,9 +591,10 @@ func TestPopDirectories(t *testing.T) {
_, err = importer.Import( _, err = importer.Import(
context.Background(), context.Background(),
f, f,
func(fileReader io.Reader) error { return nil }, func(blobReader io.Reader) ([]byte, error) { return mustBlobDigest(blobReader), nil },
func(directory *castorev1pb.Directory) error { func(directory *castorev1pb.Directory) ([]byte, error) {
return directory.Validate() require.NoError(t, directory.Validate(), "directory validation shouldn't error")
return mustDirectoryDigest(directory), nil
}, },
) )
require.NoError(t, err) require.NoError(t, err)

View file

@ -17,15 +17,15 @@ const chunkSize = 1024 * 1024
// this produces a callback function that can be used as blobCb for the // this produces a callback function that can be used as blobCb for the
// importer.Import function call. // importer.Import function call.
func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.BlobServiceClient) func(io.Reader) error { func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.BlobServiceClient) func(io.Reader) ([]byte, error) {
return func(blobReader io.Reader) error { return func(blobReader io.Reader) ([]byte, error) {
// Ensure the blobReader is buffered to at least the chunk size. // Ensure the blobReader is buffered to at least the chunk size.
blobReader = bufio.NewReaderSize(blobReader, chunkSize) blobReader = bufio.NewReaderSize(blobReader, chunkSize)
putter, err := blobServiceClient.Put(ctx) putter, err := blobServiceClient.Put(ctx)
if err != nil { if err != nil {
// return error to the importer // return error to the importer
return fmt.Errorf("error from blob service: %w", err) return nil, fmt.Errorf("error from blob service: %w", err)
} }
blobSize := 0 blobSize := 0
@ -34,7 +34,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
for { for {
n, err := blobReader.Read(chunk) n, err := blobReader.Read(chunk)
if err != nil && !errors.Is(err, io.EOF) { if err != nil && !errors.Is(err, io.EOF) {
return fmt.Errorf("unable to read from blobreader: %w", err) return nil, fmt.Errorf("unable to read from blobreader: %w", err)
} }
if n != 0 { if n != 0 {
@ -45,7 +45,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
if err := putter.Send(&castorev1pb.BlobChunk{ if err := putter.Send(&castorev1pb.BlobChunk{
Data: chunk[:n], Data: chunk[:n],
}); err != nil { }); err != nil {
return fmt.Errorf("sending blob chunk: %w", err) return nil, fmt.Errorf("sending blob chunk: %w", err)
} }
} }
@ -58,7 +58,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
resp, err := putter.CloseAndRecv() resp, err := putter.CloseAndRecv()
if err != nil { if err != nil {
return fmt.Errorf("close blob putter: %w", err) return nil, fmt.Errorf("close blob putter: %w", err)
} }
log.WithFields(log.Fields{ log.WithFields(log.Fields{
@ -66,6 +66,6 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
"blob_size": blobSize, "blob_size": blobSize,
}).Debug("uploaded blob") }).Debug("uploaded blob")
return nil return resp.GetDigest(), nil
} }
} }

View file

@ -23,10 +23,10 @@ func NewDirectoriesUploader(ctx context.Context, directoryServiceClient castorev
} }
} }
func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error { func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) ([]byte, error) {
directoryDgst, err := directory.Digest() directoryDigest, err := directory.Digest()
if err != nil { if err != nil {
return fmt.Errorf("failed calculating directory digest: %w", err) return nil, fmt.Errorf("failed calculating directory digest: %w", err)
} }
// Send the directory to the directory service // Send the directory to the directory service
@ -34,7 +34,7 @@ func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error {
if du.directoryServicePutStream == nil { if du.directoryServicePutStream == nil {
directoryServicePutStream, err := du.directoryServiceClient.Put(du.ctx) directoryServicePutStream, err := du.directoryServiceClient.Put(du.ctx)
if err != nil { if err != nil {
return fmt.Errorf("unable to initialize directory service put stream: %v", err) return nil, fmt.Errorf("unable to initialize directory service put stream: %v", err)
} }
du.directoryServicePutStream = directoryServicePutStream du.directoryServicePutStream = directoryServicePutStream
} }
@ -42,11 +42,11 @@ func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error {
// send the directory out // send the directory out
err = du.directoryServicePutStream.Send(directory) err = du.directoryServicePutStream.Send(directory)
if err != nil { if err != nil {
return fmt.Errorf("error sending directory: %w", err) return nil, fmt.Errorf("error sending directory: %w", err)
} }
log.WithField("digest", base64.StdEncoding.EncodeToString(directoryDgst)).Debug("uploaded directory") log.WithField("digest", base64.StdEncoding.EncodeToString(directoryDigest)).Debug("uploaded directory")
return nil return directoryDigest, nil
} }
// Done is called whenever we're // Done is called whenever we're

View file

@ -44,7 +44,7 @@ func registerNarPut(s *Server) {
// buffer the body by 10MiB // buffer the body by 10MiB
bufio.NewReaderSize(r.Body, 10*1024*1024), bufio.NewReaderSize(r.Body, 10*1024*1024),
genBlobServiceWriteCb(ctx, s.blobServiceClient), genBlobServiceWriteCb(ctx, s.blobServiceClient),
func(directory *castorev1pb.Directory) error { func(directory *castorev1pb.Directory) ([]byte, error) {
return directoriesUploader.Put(directory) return directoriesUploader.Put(directory)
}, },
) )

View file

@ -3,7 +3,8 @@ package writer_test
import ( import (
"bytes" "bytes"
"context" "context"
"encoding/hex" "encoding/base64"
"fmt"
"io" "io"
"os" "os"
"testing" "testing"
@ -16,7 +17,7 @@ import (
"lukechampine.com/blake3" "lukechampine.com/blake3"
) )
func mustDigest(d *castorev1pb.Directory) []byte { func mustDirectoryDigest(d *castorev1pb.Directory) []byte {
dgst, err := d.Digest() dgst, err := d.Digest()
if err != nil { if err != nil {
panic(err) panic(err)
@ -24,6 +25,15 @@ func mustDigest(d *castorev1pb.Directory) []byte {
return dgst return dgst
} }
func mustBlobDigest(r io.Reader) []byte {
hasher := blake3.New(32, nil)
_, err := io.Copy(hasher, r)
if err != nil {
panic(err)
}
return hasher.Sum([]byte{})
}
func TestSymlink(t *testing.T) { func TestSymlink(t *testing.T) {
pathInfo := &storev1pb.PathInfo{ pathInfo := &storev1pb.PathInfo{
@ -108,7 +118,7 @@ func TestEmptyDirectory(t *testing.T) {
Files: []*castorev1pb.FileNode{}, Files: []*castorev1pb.FileNode{},
Symlinks: []*castorev1pb.SymlinkNode{}, Symlinks: []*castorev1pb.SymlinkNode{},
} }
emptyDirectoryDigest := mustDigest(emptyDirectory) emptyDirectoryDigest := mustDirectoryDigest(emptyDirectory)
pathInfo := &storev1pb.PathInfo{ pathInfo := &storev1pb.PathInfo{
Node: &castorev1pb.Node{ Node: &castorev1pb.Node{
@ -156,54 +166,52 @@ func TestFull(t *testing.T) {
narContents, err := io.ReadAll(f) narContents, err := io.ReadAll(f)
require.NoError(t, err) require.NoError(t, err)
filesMap := make(map[string][]byte, 0) blobsMap := make(map[string][]byte, 0)
directoriesMap := make(map[string]*castorev1pb.Directory) directoriesMap := make(map[string]*castorev1pb.Directory)
pathInfo, err := importer.Import( pathInfo, err := importer.Import(
context.Background(), context.Background(),
bytes.NewBuffer(narContents), bytes.NewBuffer(narContents),
func(fileReader io.Reader) error { func(blobReader io.Reader) ([]byte, error) {
fileContents, err := io.ReadAll(fileReader) // read in contents, we need to put it into filesMap later.
contents, err := io.ReadAll(blobReader)
require.NoError(t, err) require.NoError(t, err)
b3Writer := blake3.New(32, nil) dgst := mustBlobDigest(bytes.NewReader(contents))
_, err = io.Copy(b3Writer, bytes.NewReader(fileContents))
require.NoError(t, err)
// put it in filesMap // put it in filesMap
filesMap[hex.EncodeToString(b3Writer.Sum(nil))] = fileContents blobsMap[base64.StdEncoding.EncodeToString(dgst)] = contents
return nil return dgst, nil
}, },
func(directory *castorev1pb.Directory) error { func(directory *castorev1pb.Directory) ([]byte, error) {
dgst := mustDigest(directory) dgst := mustDirectoryDigest(directory)
directoriesMap[hex.EncodeToString(dgst)] = directory directoriesMap[base64.StdEncoding.EncodeToString(dgst)] = directory
return nil return dgst, nil
}, },
) )
require.NoError(t, err) require.NoError(t, err)
// done populating everything, now actually test the export :-) // done populating everything, now actually test the export :-)
var buf bytes.Buffer var buf bytes.Buffer
err = writer.Export( err = writer.Export(
&buf, &buf,
pathInfo, pathInfo,
func(directoryRef []byte) (*castorev1pb.Directory, error) { func(directoryDgst []byte) (*castorev1pb.Directory, error) {
d, found := directoriesMap[hex.EncodeToString(directoryRef)] d, found := directoriesMap[base64.StdEncoding.EncodeToString(directoryDgst)]
if !found { if !found {
panic("directories not found") panic(fmt.Sprintf("directory %v not found", base64.StdEncoding.EncodeToString(directoryDgst)))
} }
return d, nil return d, nil
}, },
func(fileRef []byte) (io.ReadCloser, error) { func(blobDgst []byte) (io.ReadCloser, error) {
fileContents, found := filesMap[hex.EncodeToString(fileRef)] blobContents, found := blobsMap[base64.StdEncoding.EncodeToString(blobDgst)]
if !found { if !found {
panic("file not found") panic(fmt.Sprintf("blob %v not found", base64.StdEncoding.EncodeToString(blobDgst)))
} }
return io.NopCloser(bytes.NewReader(fileContents)), nil return io.NopCloser(bytes.NewReader(blobContents)), nil
}, },
) )