2023-10-10 20:52:03 +02:00
|
|
|
package importer_test
|
2023-10-10 20:04:19 +02:00
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"encoding/base64"
|
|
|
|
"fmt"
|
|
|
|
"io"
|
|
|
|
"os"
|
|
|
|
"testing"
|
|
|
|
|
2023-10-17 22:41:33 +02:00
|
|
|
castorev1pb "code.tvl.fyi/tvix/castore-go"
|
2023-10-10 20:04:19 +02:00
|
|
|
"code.tvl.fyi/tvix/nar-bridge/pkg/importer"
|
2023-10-17 22:41:33 +02:00
|
|
|
storev1pb "code.tvl.fyi/tvix/store-go"
|
2023-10-10 20:04:19 +02:00
|
|
|
"github.com/stretchr/testify/require"
|
|
|
|
)
|
|
|
|
|
2023-10-10 20:52:03 +02:00
|
|
|
func TestRoundtrip(t *testing.T) {
|
|
|
|
// We pipe nar_1094wph9z4nwlgvsd53abfz8i117ykiv5dwnq9nnhz846s7xqd7d.nar to
|
|
|
|
// storev1pb.Export, and store all the file contents and directory objects
|
|
|
|
// received in two hashmaps.
|
2023-10-10 20:04:19 +02:00
|
|
|
// We then feed it to the writer, and test we come up with the same NAR file.
|
|
|
|
|
|
|
|
f, err := os.Open("../../testdata/nar_1094wph9z4nwlgvsd53abfz8i117ykiv5dwnq9nnhz846s7xqd7d.nar")
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
narContents, err := io.ReadAll(f)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
blobsMap := make(map[string][]byte, 0)
|
|
|
|
directoriesMap := make(map[string]*castorev1pb.Directory)
|
|
|
|
|
2023-10-11 12:28:10 +02:00
|
|
|
rootNode, _, _, err := importer.Import(
|
2023-10-10 20:04:19 +02:00
|
|
|
context.Background(),
|
|
|
|
bytes.NewBuffer(narContents),
|
|
|
|
func(blobReader io.Reader) ([]byte, error) {
|
|
|
|
// read in contents, we need to put it into filesMap later.
|
|
|
|
contents, err := io.ReadAll(blobReader)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
dgst := mustBlobDigest(bytes.NewReader(contents))
|
|
|
|
|
|
|
|
// put it in filesMap
|
|
|
|
blobsMap[base64.StdEncoding.EncodeToString(dgst)] = contents
|
|
|
|
|
|
|
|
return dgst, nil
|
|
|
|
},
|
|
|
|
func(directory *castorev1pb.Directory) ([]byte, error) {
|
|
|
|
dgst := mustDirectoryDigest(directory)
|
|
|
|
|
|
|
|
directoriesMap[base64.StdEncoding.EncodeToString(dgst)] = directory
|
|
|
|
return dgst, nil
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// done populating everything, now actually test the export :-)
|
2023-10-11 12:28:10 +02:00
|
|
|
var narBuf bytes.Buffer
|
2023-10-10 20:52:03 +02:00
|
|
|
err = storev1pb.Export(
|
2023-10-11 12:28:10 +02:00
|
|
|
&narBuf,
|
|
|
|
rootNode,
|
2023-10-10 20:04:19 +02:00
|
|
|
func(directoryDgst []byte) (*castorev1pb.Directory, error) {
|
|
|
|
d, found := directoriesMap[base64.StdEncoding.EncodeToString(directoryDgst)]
|
|
|
|
if !found {
|
|
|
|
panic(fmt.Sprintf("directory %v not found", base64.StdEncoding.EncodeToString(directoryDgst)))
|
|
|
|
}
|
|
|
|
return d, nil
|
|
|
|
},
|
|
|
|
func(blobDgst []byte) (io.ReadCloser, error) {
|
|
|
|
blobContents, found := blobsMap[base64.StdEncoding.EncodeToString(blobDgst)]
|
|
|
|
if !found {
|
|
|
|
panic(fmt.Sprintf("blob %v not found", base64.StdEncoding.EncodeToString(blobDgst)))
|
|
|
|
}
|
|
|
|
return io.NopCloser(bytes.NewReader(blobContents)), nil
|
|
|
|
},
|
|
|
|
)
|
|
|
|
|
|
|
|
require.NoError(t, err, "exporter shouldn't fail")
|
2023-10-11 12:28:10 +02:00
|
|
|
require.Equal(t, narContents, narBuf.Bytes())
|
2023-10-10 20:04:19 +02:00
|
|
|
}
|