refactor(tvix/nar-bridge): let callbaks return calculated digests
This aligns behaviour more with how it should be - it's the responsibility of the callback functions to return digests of the things they consume(d). It allows further cleaning up the hasher struct. Change-Id: I9cbfc87e6abd4ff17fadf39eb6563ec3cb7fcc6f Reviewed-on: https://cl.tvl.fyi/c/depot/+/9528 Autosubmit: flokli <flokli@flokli.de> Tested-by: BuildkiteCI Reviewed-by: Connor Brewster <cbrewster@hey.com>
This commit is contained in:
		
							parent
							
								
									b1ff1267be
								
							
						
					
					
						commit
						f92b0ef933
					
				
					 6 changed files with 109 additions and 100 deletions
				
			
		| 
						 | 
					@ -30,9 +30,9 @@ func Import(
 | 
				
			||||||
	// The reader the data is read from
 | 
						// The reader the data is read from
 | 
				
			||||||
	r io.Reader,
 | 
						r io.Reader,
 | 
				
			||||||
	// callback function called with each regular file content
 | 
						// callback function called with each regular file content
 | 
				
			||||||
	blobCb func(fileReader io.Reader) error,
 | 
						blobCb func(fileReader io.Reader) ([]byte, error),
 | 
				
			||||||
	// callback function called with each finalized directory node
 | 
						// callback function called with each finalized directory node
 | 
				
			||||||
	directoryCb func(directory *castorev1pb.Directory) error,
 | 
						directoryCb func(directory *castorev1pb.Directory) ([]byte, error),
 | 
				
			||||||
) (*storev1pb.PathInfo, error) {
 | 
					) (*storev1pb.PathInfo, error) {
 | 
				
			||||||
	// wrap the passed reader in a reader that records the number of bytes read and
 | 
						// wrap the passed reader in a reader that records the number of bytes read and
 | 
				
			||||||
	// their sha256 sum.
 | 
						// their sha256 sum.
 | 
				
			||||||
| 
						 | 
					@ -65,24 +65,21 @@ func Import(
 | 
				
			||||||
		toPop := stack[len(stack)-1]
 | 
							toPop := stack[len(stack)-1]
 | 
				
			||||||
		stack = stack[:len(stack)-1]
 | 
							stack = stack[:len(stack)-1]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							// call the directoryCb
 | 
				
			||||||
 | 
							directoryDigest, err := directoryCb(toPop.directory)
 | 
				
			||||||
 | 
							if err != nil {
 | 
				
			||||||
 | 
								return fmt.Errorf("failed calling directoryCb: %w", err)
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		// if there's still a parent left on the stack, refer to it from there.
 | 
							// if there's still a parent left on the stack, refer to it from there.
 | 
				
			||||||
		if len(stack) > 0 {
 | 
							if len(stack) > 0 {
 | 
				
			||||||
			dgst, err := toPop.directory.Digest()
 | 
					 | 
				
			||||||
			if err != nil {
 | 
					 | 
				
			||||||
				return fmt.Errorf("unable to calculate directory digest: %w", err)
 | 
					 | 
				
			||||||
			}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
			topOfStack := stack[len(stack)-1].directory
 | 
								topOfStack := stack[len(stack)-1].directory
 | 
				
			||||||
			topOfStack.Directories = append(topOfStack.Directories, &castorev1pb.DirectoryNode{
 | 
								topOfStack.Directories = append(topOfStack.Directories, &castorev1pb.DirectoryNode{
 | 
				
			||||||
				Name:   []byte(path.Base(toPop.path)),
 | 
									Name:   []byte(path.Base(toPop.path)),
 | 
				
			||||||
				Digest: dgst,
 | 
									Digest: directoryDigest,
 | 
				
			||||||
				Size:   toPop.directory.Size(),
 | 
									Size:   toPop.directory.Size(),
 | 
				
			||||||
			})
 | 
								})
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		// call the directoryCb
 | 
					 | 
				
			||||||
		if err := directoryCb(toPop.directory); err != nil {
 | 
					 | 
				
			||||||
			return fmt.Errorf("failed calling directoryCb: %w", err)
 | 
					 | 
				
			||||||
		}
 | 
					 | 
				
			||||||
		// Keep track that we have encounter at least one directory
 | 
							// Keep track that we have encounter at least one directory
 | 
				
			||||||
		stackDirectory = toPop.directory
 | 
							stackDirectory = toPop.directory
 | 
				
			||||||
		return nil
 | 
							return nil
 | 
				
			||||||
| 
						 | 
					@ -106,7 +103,7 @@ func Import(
 | 
				
			||||||
			hdr, err := narReader.Next()
 | 
								hdr, err := narReader.Next()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			// If this returns an error, it's either EOF (when we're done reading from the NAR),
 | 
								// If this returns an error, it's either EOF (when we're done reading from the NAR),
 | 
				
			||||||
			// or another error
 | 
								// or another error.
 | 
				
			||||||
			if err != nil {
 | 
								if err != nil {
 | 
				
			||||||
				// if this returns no EOF, bail out
 | 
									// if this returns no EOF, bail out
 | 
				
			||||||
				if !errors.Is(err, io.EOF) {
 | 
									if !errors.Is(err, io.EOF) {
 | 
				
			||||||
| 
						 | 
					@ -206,28 +203,22 @@ func Import(
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
			if hdr.Type == nar.TypeRegular {
 | 
								if hdr.Type == nar.TypeRegular {
 | 
				
			||||||
				// wrap reader with a reader calculating the blake3 hash
 | 
									// wrap reader with a reader calculating the blake3 hash
 | 
				
			||||||
				fileReader := hashers.NewHasher(narReader, blake3.New(32, nil))
 | 
									blobReader := hashers.NewHasher(narReader, blake3.New(32, nil))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
				err := blobCb(fileReader)
 | 
									blobDigest, err := blobCb(blobReader)
 | 
				
			||||||
				if err != nil {
 | 
									if err != nil {
 | 
				
			||||||
					return nil, fmt.Errorf("failure from blobCb: %w", err)
 | 
										return nil, fmt.Errorf("failure from blobCb: %w", err)
 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
				// drive the file reader to the end, in case the CB function doesn't read
 | 
									// ensure blobCb did read all the way to the end.
 | 
				
			||||||
				// all the way to the end on its own
 | 
									// If it didn't, the blobCb function is wrong and we should bail out.
 | 
				
			||||||
				if fileReader.BytesWritten() != uint32(hdr.Size) {
 | 
									if blobReader.BytesWritten() != uint32(hdr.Size) {
 | 
				
			||||||
					_, err := io.ReadAll(fileReader)
 | 
										panic("not read to end")
 | 
				
			||||||
					if err != nil {
 | 
					 | 
				
			||||||
						return nil, fmt.Errorf("unable to read until the end of the file content: %w", err)
 | 
					 | 
				
			||||||
					}
 | 
					 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
				// read the blake3 hash
 | 
					 | 
				
			||||||
				dgst := fileReader.Sum(nil)
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
				fileNode := &castorev1pb.FileNode{
 | 
									fileNode := &castorev1pb.FileNode{
 | 
				
			||||||
					Name:       []byte(getBasename(hdr.Path)),
 | 
										Name:       []byte(getBasename(hdr.Path)),
 | 
				
			||||||
					Digest:     dgst,
 | 
										Digest:     blobDigest,
 | 
				
			||||||
					Size:       uint32(hdr.Size),
 | 
										Size:       uint32(hdr.Size),
 | 
				
			||||||
					Executable: hdr.Executable,
 | 
										Executable: hdr.Executable,
 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1,6 +1,7 @@
 | 
				
			||||||
package importer_test
 | 
					package importer_test
 | 
				
			||||||
 | 
					
 | 
				
			||||||
import (
 | 
					import (
 | 
				
			||||||
 | 
						"bytes"
 | 
				
			||||||
	"context"
 | 
						"context"
 | 
				
			||||||
	"errors"
 | 
						"errors"
 | 
				
			||||||
	"io"
 | 
						"io"
 | 
				
			||||||
| 
						 | 
					@ -13,6 +14,7 @@ import (
 | 
				
			||||||
	"github.com/google/go-cmp/cmp"
 | 
						"github.com/google/go-cmp/cmp"
 | 
				
			||||||
	"github.com/stretchr/testify/require"
 | 
						"github.com/stretchr/testify/require"
 | 
				
			||||||
	"google.golang.org/protobuf/testing/protocmp"
 | 
						"google.golang.org/protobuf/testing/protocmp"
 | 
				
			||||||
 | 
						"lukechampine.com/blake3"
 | 
				
			||||||
)
 | 
					)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) {
 | 
					func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) {
 | 
				
			||||||
| 
						 | 
					@ -21,7 +23,7 @@ func requireProtoEq(t *testing.T, expected interface{}, actual interface{}) {
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func mustDigest(d *castorev1pb.Directory) []byte {
 | 
					func mustDirectoryDigest(d *castorev1pb.Directory) []byte {
 | 
				
			||||||
	dgst, err := d.Digest()
 | 
						dgst, err := d.Digest()
 | 
				
			||||||
	if err != nil {
 | 
						if err != nil {
 | 
				
			||||||
		panic(err)
 | 
							panic(err)
 | 
				
			||||||
| 
						 | 
					@ -29,6 +31,15 @@ func mustDigest(d *castorev1pb.Directory) []byte {
 | 
				
			||||||
	return dgst
 | 
						return dgst
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					func mustBlobDigest(r io.Reader) []byte {
 | 
				
			||||||
 | 
						hasher := blake3.New(32, nil)
 | 
				
			||||||
 | 
						_, err := io.Copy(hasher, r)
 | 
				
			||||||
 | 
						if err != nil {
 | 
				
			||||||
 | 
							panic(err)
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						return hasher.Sum([]byte{})
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func TestSymlink(t *testing.T) {
 | 
					func TestSymlink(t *testing.T) {
 | 
				
			||||||
	f, err := os.Open("../../testdata/symlink.nar")
 | 
						f, err := os.Open("../../testdata/symlink.nar")
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
| 
						 | 
					@ -36,9 +47,9 @@ func TestSymlink(t *testing.T) {
 | 
				
			||||||
	actualPathInfo, err := importer.Import(
 | 
						actualPathInfo, err := importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		f,
 | 
							f,
 | 
				
			||||||
		func(fileReader io.Reader) error {
 | 
							func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
			panic("no file contents expected!")
 | 
								panic("no file contents expected!")
 | 
				
			||||||
		}, func(directory *castorev1pb.Directory) error {
 | 
							}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			panic("no directories expected!")
 | 
								panic("no directories expected!")
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
| 
						 | 
					@ -74,12 +85,12 @@ func TestRegular(t *testing.T) {
 | 
				
			||||||
	actualPathInfo, err := importer.Import(
 | 
						actualPathInfo, err := importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		f,
 | 
							f,
 | 
				
			||||||
		func(fileReader io.Reader) error {
 | 
							func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
			contents, err := io.ReadAll(fileReader)
 | 
								contents, err := io.ReadAll(blobReader)
 | 
				
			||||||
			require.NoError(t, err, "reading fileReader should not error")
 | 
								require.NoError(t, err, "reading blobReader should not error")
 | 
				
			||||||
			require.Equal(t, []byte{0x01}, contents, "contents read from fileReader should match expectations")
 | 
								require.Equal(t, []byte{0x01}, contents, "contents read from blobReader should match expectations")
 | 
				
			||||||
			return nil
 | 
								return mustBlobDigest(bytes.NewBuffer(contents)), nil
 | 
				
			||||||
		}, func(directory *castorev1pb.Directory) error {
 | 
							}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			panic("no directories expected!")
 | 
								panic("no directories expected!")
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
| 
						 | 
					@ -129,11 +140,11 @@ func TestEmptyDirectory(t *testing.T) {
 | 
				
			||||||
	actualPathInfo, err := importer.Import(
 | 
						actualPathInfo, err := importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		f,
 | 
							f,
 | 
				
			||||||
		func(fileReader io.Reader) error {
 | 
							func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
			panic("no file contents expected!")
 | 
								panic("no file contents expected!")
 | 
				
			||||||
		}, func(directory *castorev1pb.Directory) error {
 | 
							}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			requireProtoEq(t, expectedDirectory, directory)
 | 
								requireProtoEq(t, expectedDirectory, directory)
 | 
				
			||||||
			return nil
 | 
								return mustDirectoryDigest(directory), nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
| 
						 | 
					@ -143,7 +154,7 @@ func TestEmptyDirectory(t *testing.T) {
 | 
				
			||||||
			Node: &castorev1pb.Node_Directory{
 | 
								Node: &castorev1pb.Node_Directory{
 | 
				
			||||||
				Directory: &castorev1pb.DirectoryNode{
 | 
									Directory: &castorev1pb.DirectoryNode{
 | 
				
			||||||
					Name:   []byte(""),
 | 
										Name:   []byte(""),
 | 
				
			||||||
					Digest: mustDigest(expectedDirectory),
 | 
										Digest: mustDirectoryDigest(expectedDirectory),
 | 
				
			||||||
					Size:   expectedDirectory.Size(),
 | 
										Size:   expectedDirectory.Size(),
 | 
				
			||||||
				},
 | 
									},
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
| 
						 | 
					@ -415,17 +426,17 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
		Directories: []*castorev1pb.DirectoryNode{
 | 
							Directories: []*castorev1pb.DirectoryNode{
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("man1"),
 | 
									Name:   []byte("man1"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/share/man/man1"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/share/man/man1"]),
 | 
				
			||||||
				Size:   expectedDirectories["/share/man/man1"].Size(),
 | 
									Size:   expectedDirectories["/share/man/man1"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("man5"),
 | 
									Name:   []byte("man5"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/share/man/man5"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/share/man/man5"]),
 | 
				
			||||||
				Size:   expectedDirectories["/share/man/man5"].Size(),
 | 
									Size:   expectedDirectories["/share/man/man5"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("man8"),
 | 
									Name:   []byte("man8"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/share/man/man8"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/share/man/man8"]),
 | 
				
			||||||
				Size:   expectedDirectories["/share/man/man8"].Size(),
 | 
									Size:   expectedDirectories["/share/man/man8"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
| 
						 | 
					@ -438,7 +449,7 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
		Directories: []*castorev1pb.DirectoryNode{
 | 
							Directories: []*castorev1pb.DirectoryNode{
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("man"),
 | 
									Name:   []byte("man"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/share/man"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/share/man"]),
 | 
				
			||||||
				Size:   expectedDirectories["/share/man"].Size(),
 | 
									Size:   expectedDirectories["/share/man"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
| 
						 | 
					@ -451,12 +462,12 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
		Directories: []*castorev1pb.DirectoryNode{
 | 
							Directories: []*castorev1pb.DirectoryNode{
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("bin"),
 | 
									Name:   []byte("bin"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/bin"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/bin"]),
 | 
				
			||||||
				Size:   expectedDirectories["/bin"].Size(),
 | 
									Size:   expectedDirectories["/bin"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
			{
 | 
								{
 | 
				
			||||||
				Name:   []byte("share"),
 | 
									Name:   []byte("share"),
 | 
				
			||||||
				Digest: mustDigest(expectedDirectories["/share"]),
 | 
									Digest: mustDirectoryDigest(expectedDirectories["/share"]),
 | 
				
			||||||
				Size:   expectedDirectories["/share"].Size(),
 | 
									Size:   expectedDirectories["/share"].Size(),
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
| 
						 | 
					@ -476,14 +487,12 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
	actualPathInfo, err := importer.Import(
 | 
						actualPathInfo, err := importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		f,
 | 
							f,
 | 
				
			||||||
		func(fileReader io.Reader) error {
 | 
							func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
			// Don't really bother reading and comparing the contents here,
 | 
								// Don't really bother reading and comparing the contents here,
 | 
				
			||||||
			// We already verify the right digests are produced by comparing the
 | 
								// We already verify the right digests are produced by comparing the
 | 
				
			||||||
			// directoryCb calls, and TestRegular ensures the reader works.
 | 
								// directoryCb calls, and TestRegular ensures the reader works.
 | 
				
			||||||
			// This also covers the case when the client doesn't read from the reader, and that the
 | 
								return mustBlobDigest(blobReader), nil
 | 
				
			||||||
			// importer will take care of reading all the way to the end no matter what.
 | 
							}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			return nil
 | 
					 | 
				
			||||||
		}, func(directory *castorev1pb.Directory) error {
 | 
					 | 
				
			||||||
			// use actualDirectoryOrder to look up the Directory object we expect at this specific invocation.
 | 
								// use actualDirectoryOrder to look up the Directory object we expect at this specific invocation.
 | 
				
			||||||
			currentDirectoryPath := expectedDirectoryPaths[numDirectoriesReceived]
 | 
								currentDirectoryPath := expectedDirectoryPaths[numDirectoriesReceived]
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -493,7 +502,7 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
			requireProtoEq(t, expectedDirectory, directory)
 | 
								requireProtoEq(t, expectedDirectory, directory)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			numDirectoriesReceived += 1
 | 
								numDirectoriesReceived += 1
 | 
				
			||||||
			return nil
 | 
								return mustDirectoryDigest(directory), nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
| 
						 | 
					@ -503,7 +512,7 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
			Node: &castorev1pb.Node_Directory{
 | 
								Node: &castorev1pb.Node_Directory{
 | 
				
			||||||
				Directory: &castorev1pb.DirectoryNode{
 | 
									Directory: &castorev1pb.DirectoryNode{
 | 
				
			||||||
					Name:   []byte(""),
 | 
										Name:   []byte(""),
 | 
				
			||||||
					Digest: mustDigest(expectedDirectories["/"]),
 | 
										Digest: mustDirectoryDigest(expectedDirectories["/"]),
 | 
				
			||||||
					Size:   expectedDirectories["/"].Size(),
 | 
										Size:   expectedDirectories["/"].Size(),
 | 
				
			||||||
				},
 | 
									},
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
| 
						 | 
					@ -524,7 +533,7 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
// TestCallbackErrors ensures that errors returned from the callback function
 | 
					// TestCallbackErrors ensures that errors returned from the callback function
 | 
				
			||||||
// bubble up to the importer process, and are not ignored.
 | 
					// bubble up to the importer process, and are not ignored.
 | 
				
			||||||
func TestCallbackErrors(t *testing.T) {
 | 
					func TestCallbackErrors(t *testing.T) {
 | 
				
			||||||
	t.Run("callback file", func(t *testing.T) {
 | 
						t.Run("callback blob", func(t *testing.T) {
 | 
				
			||||||
		// Pick an example NAR with a regular file.
 | 
							// Pick an example NAR with a regular file.
 | 
				
			||||||
		f, err := os.Open("../../testdata/onebyteregular.nar")
 | 
							f, err := os.Open("../../testdata/onebyteregular.nar")
 | 
				
			||||||
		require.NoError(t, err)
 | 
							require.NoError(t, err)
 | 
				
			||||||
| 
						 | 
					@ -534,9 +543,9 @@ func TestCallbackErrors(t *testing.T) {
 | 
				
			||||||
		_, err = importer.Import(
 | 
							_, err = importer.Import(
 | 
				
			||||||
			context.Background(),
 | 
								context.Background(),
 | 
				
			||||||
			f,
 | 
								f,
 | 
				
			||||||
			func(fileReader io.Reader) error {
 | 
								func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
				return targetErr
 | 
									return nil, targetErr
 | 
				
			||||||
			}, func(directory *castorev1pb.Directory) error {
 | 
								}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
				panic("no directories expected!")
 | 
									panic("no directories expected!")
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		)
 | 
							)
 | 
				
			||||||
| 
						 | 
					@ -552,10 +561,10 @@ func TestCallbackErrors(t *testing.T) {
 | 
				
			||||||
		_, err = importer.Import(
 | 
							_, err = importer.Import(
 | 
				
			||||||
			context.Background(),
 | 
								context.Background(),
 | 
				
			||||||
			f,
 | 
								f,
 | 
				
			||||||
			func(fileReader io.Reader) error {
 | 
								func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
				panic("no file contents expected!")
 | 
									panic("no file contents expected!")
 | 
				
			||||||
			}, func(directory *castorev1pb.Directory) error {
 | 
								}, func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
				return targetErr
 | 
									return nil, targetErr
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		)
 | 
							)
 | 
				
			||||||
		require.ErrorIs(t, err, targetErr)
 | 
							require.ErrorIs(t, err, targetErr)
 | 
				
			||||||
| 
						 | 
					@ -582,9 +591,10 @@ func TestPopDirectories(t *testing.T) {
 | 
				
			||||||
	_, err = importer.Import(
 | 
						_, err = importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		f,
 | 
							f,
 | 
				
			||||||
		func(fileReader io.Reader) error { return nil },
 | 
							func(blobReader io.Reader) ([]byte, error) { return mustBlobDigest(blobReader), nil },
 | 
				
			||||||
		func(directory *castorev1pb.Directory) error {
 | 
							func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			return directory.Validate()
 | 
								require.NoError(t, directory.Validate(), "directory validation shouldn't error")
 | 
				
			||||||
 | 
								return mustDirectoryDigest(directory), nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -17,15 +17,15 @@ const chunkSize = 1024 * 1024
 | 
				
			||||||
 | 
					
 | 
				
			||||||
// this produces a callback function that can be used as blobCb for the
 | 
					// this produces a callback function that can be used as blobCb for the
 | 
				
			||||||
// importer.Import function call.
 | 
					// importer.Import function call.
 | 
				
			||||||
func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.BlobServiceClient) func(io.Reader) error {
 | 
					func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.BlobServiceClient) func(io.Reader) ([]byte, error) {
 | 
				
			||||||
	return func(blobReader io.Reader) error {
 | 
						return func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
		// Ensure the blobReader is buffered to at least the chunk size.
 | 
							// Ensure the blobReader is buffered to at least the chunk size.
 | 
				
			||||||
		blobReader = bufio.NewReaderSize(blobReader, chunkSize)
 | 
							blobReader = bufio.NewReaderSize(blobReader, chunkSize)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		putter, err := blobServiceClient.Put(ctx)
 | 
							putter, err := blobServiceClient.Put(ctx)
 | 
				
			||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			// return error to the importer
 | 
								// return error to the importer
 | 
				
			||||||
			return fmt.Errorf("error from blob service: %w", err)
 | 
								return nil, fmt.Errorf("error from blob service: %w", err)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		blobSize := 0
 | 
							blobSize := 0
 | 
				
			||||||
| 
						 | 
					@ -34,7 +34,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
 | 
				
			||||||
		for {
 | 
							for {
 | 
				
			||||||
			n, err := blobReader.Read(chunk)
 | 
								n, err := blobReader.Read(chunk)
 | 
				
			||||||
			if err != nil && !errors.Is(err, io.EOF) {
 | 
								if err != nil && !errors.Is(err, io.EOF) {
 | 
				
			||||||
				return fmt.Errorf("unable to read from blobreader: %w", err)
 | 
									return nil, fmt.Errorf("unable to read from blobreader: %w", err)
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			if n != 0 {
 | 
								if n != 0 {
 | 
				
			||||||
| 
						 | 
					@ -45,7 +45,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
 | 
				
			||||||
				if err := putter.Send(&castorev1pb.BlobChunk{
 | 
									if err := putter.Send(&castorev1pb.BlobChunk{
 | 
				
			||||||
					Data: chunk[:n],
 | 
										Data: chunk[:n],
 | 
				
			||||||
				}); err != nil {
 | 
									}); err != nil {
 | 
				
			||||||
					return fmt.Errorf("sending blob chunk: %w", err)
 | 
										return nil, fmt.Errorf("sending blob chunk: %w", err)
 | 
				
			||||||
				}
 | 
									}
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -58,7 +58,7 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		resp, err := putter.CloseAndRecv()
 | 
							resp, err := putter.CloseAndRecv()
 | 
				
			||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			return fmt.Errorf("close blob putter: %w", err)
 | 
								return nil, fmt.Errorf("close blob putter: %w", err)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		log.WithFields(log.Fields{
 | 
							log.WithFields(log.Fields{
 | 
				
			||||||
| 
						 | 
					@ -66,6 +66,6 @@ func genBlobServiceWriteCb(ctx context.Context, blobServiceClient castorev1pb.Bl
 | 
				
			||||||
			"blob_size":   blobSize,
 | 
								"blob_size":   blobSize,
 | 
				
			||||||
		}).Debug("uploaded blob")
 | 
							}).Debug("uploaded blob")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		return nil
 | 
							return resp.GetDigest(), nil
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -23,10 +23,10 @@ func NewDirectoriesUploader(ctx context.Context, directoryServiceClient castorev
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error {
 | 
					func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
	directoryDgst, err := directory.Digest()
 | 
						directoryDigest, err := directory.Digest()
 | 
				
			||||||
	if err != nil {
 | 
						if err != nil {
 | 
				
			||||||
		return fmt.Errorf("failed calculating directory digest: %w", err)
 | 
							return nil, fmt.Errorf("failed calculating directory digest: %w", err)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	// Send the directory to the directory service
 | 
						// Send the directory to the directory service
 | 
				
			||||||
| 
						 | 
					@ -34,7 +34,7 @@ func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error {
 | 
				
			||||||
	if du.directoryServicePutStream == nil {
 | 
						if du.directoryServicePutStream == nil {
 | 
				
			||||||
		directoryServicePutStream, err := du.directoryServiceClient.Put(du.ctx)
 | 
							directoryServicePutStream, err := du.directoryServiceClient.Put(du.ctx)
 | 
				
			||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			return fmt.Errorf("unable to initialize directory service put stream: %v", err)
 | 
								return nil, fmt.Errorf("unable to initialize directory service put stream: %v", err)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		du.directoryServicePutStream = directoryServicePutStream
 | 
							du.directoryServicePutStream = directoryServicePutStream
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
| 
						 | 
					@ -42,11 +42,11 @@ func (du *DirectoriesUploader) Put(directory *castorev1pb.Directory) error {
 | 
				
			||||||
	// send the directory out
 | 
						// send the directory out
 | 
				
			||||||
	err = du.directoryServicePutStream.Send(directory)
 | 
						err = du.directoryServicePutStream.Send(directory)
 | 
				
			||||||
	if err != nil {
 | 
						if err != nil {
 | 
				
			||||||
		return fmt.Errorf("error sending directory: %w", err)
 | 
							return nil, fmt.Errorf("error sending directory: %w", err)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	log.WithField("digest", base64.StdEncoding.EncodeToString(directoryDgst)).Debug("uploaded directory")
 | 
						log.WithField("digest", base64.StdEncoding.EncodeToString(directoryDigest)).Debug("uploaded directory")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return nil
 | 
						return directoryDigest, nil
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
// Done is called whenever we're
 | 
					// Done is called whenever we're
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -44,7 +44,7 @@ func registerNarPut(s *Server) {
 | 
				
			||||||
			// buffer the body by 10MiB
 | 
								// buffer the body by 10MiB
 | 
				
			||||||
			bufio.NewReaderSize(r.Body, 10*1024*1024),
 | 
								bufio.NewReaderSize(r.Body, 10*1024*1024),
 | 
				
			||||||
			genBlobServiceWriteCb(ctx, s.blobServiceClient),
 | 
								genBlobServiceWriteCb(ctx, s.blobServiceClient),
 | 
				
			||||||
			func(directory *castorev1pb.Directory) error {
 | 
								func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
				return directoriesUploader.Put(directory)
 | 
									return directoriesUploader.Put(directory)
 | 
				
			||||||
			},
 | 
								},
 | 
				
			||||||
		)
 | 
							)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3,7 +3,8 @@ package writer_test
 | 
				
			||||||
import (
 | 
					import (
 | 
				
			||||||
	"bytes"
 | 
						"bytes"
 | 
				
			||||||
	"context"
 | 
						"context"
 | 
				
			||||||
	"encoding/hex"
 | 
						"encoding/base64"
 | 
				
			||||||
 | 
						"fmt"
 | 
				
			||||||
	"io"
 | 
						"io"
 | 
				
			||||||
	"os"
 | 
						"os"
 | 
				
			||||||
	"testing"
 | 
						"testing"
 | 
				
			||||||
| 
						 | 
					@ -16,7 +17,7 @@ import (
 | 
				
			||||||
	"lukechampine.com/blake3"
 | 
						"lukechampine.com/blake3"
 | 
				
			||||||
)
 | 
					)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func mustDigest(d *castorev1pb.Directory) []byte {
 | 
					func mustDirectoryDigest(d *castorev1pb.Directory) []byte {
 | 
				
			||||||
	dgst, err := d.Digest()
 | 
						dgst, err := d.Digest()
 | 
				
			||||||
	if err != nil {
 | 
						if err != nil {
 | 
				
			||||||
		panic(err)
 | 
							panic(err)
 | 
				
			||||||
| 
						 | 
					@ -24,6 +25,15 @@ func mustDigest(d *castorev1pb.Directory) []byte {
 | 
				
			||||||
	return dgst
 | 
						return dgst
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					func mustBlobDigest(r io.Reader) []byte {
 | 
				
			||||||
 | 
						hasher := blake3.New(32, nil)
 | 
				
			||||||
 | 
						_, err := io.Copy(hasher, r)
 | 
				
			||||||
 | 
						if err != nil {
 | 
				
			||||||
 | 
							panic(err)
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						return hasher.Sum([]byte{})
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func TestSymlink(t *testing.T) {
 | 
					func TestSymlink(t *testing.T) {
 | 
				
			||||||
	pathInfo := &storev1pb.PathInfo{
 | 
						pathInfo := &storev1pb.PathInfo{
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -108,7 +118,7 @@ func TestEmptyDirectory(t *testing.T) {
 | 
				
			||||||
		Files:       []*castorev1pb.FileNode{},
 | 
							Files:       []*castorev1pb.FileNode{},
 | 
				
			||||||
		Symlinks:    []*castorev1pb.SymlinkNode{},
 | 
							Symlinks:    []*castorev1pb.SymlinkNode{},
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	emptyDirectoryDigest := mustDigest(emptyDirectory)
 | 
						emptyDirectoryDigest := mustDirectoryDigest(emptyDirectory)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	pathInfo := &storev1pb.PathInfo{
 | 
						pathInfo := &storev1pb.PathInfo{
 | 
				
			||||||
		Node: &castorev1pb.Node{
 | 
							Node: &castorev1pb.Node{
 | 
				
			||||||
| 
						 | 
					@ -156,54 +166,52 @@ func TestFull(t *testing.T) {
 | 
				
			||||||
	narContents, err := io.ReadAll(f)
 | 
						narContents, err := io.ReadAll(f)
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	filesMap := make(map[string][]byte, 0)
 | 
						blobsMap := make(map[string][]byte, 0)
 | 
				
			||||||
	directoriesMap := make(map[string]*castorev1pb.Directory)
 | 
						directoriesMap := make(map[string]*castorev1pb.Directory)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	pathInfo, err := importer.Import(
 | 
						pathInfo, err := importer.Import(
 | 
				
			||||||
		context.Background(),
 | 
							context.Background(),
 | 
				
			||||||
		bytes.NewBuffer(narContents),
 | 
							bytes.NewBuffer(narContents),
 | 
				
			||||||
		func(fileReader io.Reader) error {
 | 
							func(blobReader io.Reader) ([]byte, error) {
 | 
				
			||||||
			fileContents, err := io.ReadAll(fileReader)
 | 
								// read in contents, we need to put it into filesMap later.
 | 
				
			||||||
 | 
								contents, err := io.ReadAll(blobReader)
 | 
				
			||||||
			require.NoError(t, err)
 | 
								require.NoError(t, err)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			b3Writer := blake3.New(32, nil)
 | 
								dgst := mustBlobDigest(bytes.NewReader(contents))
 | 
				
			||||||
			_, err = io.Copy(b3Writer, bytes.NewReader(fileContents))
 | 
					 | 
				
			||||||
			require.NoError(t, err)
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
			// put it in filesMap
 | 
								// put it in filesMap
 | 
				
			||||||
			filesMap[hex.EncodeToString(b3Writer.Sum(nil))] = fileContents
 | 
								blobsMap[base64.StdEncoding.EncodeToString(dgst)] = contents
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			return nil
 | 
								return dgst, nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
		func(directory *castorev1pb.Directory) error {
 | 
							func(directory *castorev1pb.Directory) ([]byte, error) {
 | 
				
			||||||
			dgst := mustDigest(directory)
 | 
								dgst := mustDirectoryDigest(directory)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
			directoriesMap[hex.EncodeToString(dgst)] = directory
 | 
								directoriesMap[base64.StdEncoding.EncodeToString(dgst)] = directory
 | 
				
			||||||
			return nil
 | 
								return dgst, nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	require.NoError(t, err)
 | 
						require.NoError(t, err)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	// done populating everything, now actually test the export :-)
 | 
						// done populating everything, now actually test the export :-)
 | 
				
			||||||
 | 
					 | 
				
			||||||
	var buf bytes.Buffer
 | 
						var buf bytes.Buffer
 | 
				
			||||||
	err = writer.Export(
 | 
						err = writer.Export(
 | 
				
			||||||
		&buf,
 | 
							&buf,
 | 
				
			||||||
		pathInfo,
 | 
							pathInfo,
 | 
				
			||||||
		func(directoryRef []byte) (*castorev1pb.Directory, error) {
 | 
							func(directoryDgst []byte) (*castorev1pb.Directory, error) {
 | 
				
			||||||
			d, found := directoriesMap[hex.EncodeToString(directoryRef)]
 | 
								d, found := directoriesMap[base64.StdEncoding.EncodeToString(directoryDgst)]
 | 
				
			||||||
			if !found {
 | 
								if !found {
 | 
				
			||||||
				panic("directories not found")
 | 
									panic(fmt.Sprintf("directory %v not found", base64.StdEncoding.EncodeToString(directoryDgst)))
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
			return d, nil
 | 
								return d, nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
		func(fileRef []byte) (io.ReadCloser, error) {
 | 
							func(blobDgst []byte) (io.ReadCloser, error) {
 | 
				
			||||||
			fileContents, found := filesMap[hex.EncodeToString(fileRef)]
 | 
								blobContents, found := blobsMap[base64.StdEncoding.EncodeToString(blobDgst)]
 | 
				
			||||||
			if !found {
 | 
								if !found {
 | 
				
			||||||
				panic("file not found")
 | 
									panic(fmt.Sprintf("blob      %v not found", base64.StdEncoding.EncodeToString(blobDgst)))
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
			return io.NopCloser(bytes.NewReader(fileContents)), nil
 | 
								return io.NopCloser(bytes.NewReader(blobContents)), nil
 | 
				
			||||||
		},
 | 
							},
 | 
				
			||||||
	)
 | 
						)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue