This otherwise gets a bit spammy. Change-Id: I288350a600d79a394c239f253424ad55bc3cefc5 Reviewed-on: https://cl.tvl.fyi/c/depot/+/10954 Tested-by: BuildkiteCI Reviewed-by: tazjin <tazjin@tvl.su>
385 lines
13 KiB
Rust
385 lines
13 KiB
Rust
use crate::blobservice::BlobService;
|
|
use crate::directoryservice::DirectoryPutter;
|
|
use crate::directoryservice::DirectoryService;
|
|
use crate::proto::node::Node;
|
|
use crate::proto::Directory;
|
|
use crate::proto::DirectoryNode;
|
|
use crate::proto::FileNode;
|
|
use crate::proto::SymlinkNode;
|
|
use crate::Error as CastoreError;
|
|
use async_stream::stream;
|
|
use futures::pin_mut;
|
|
use futures::Stream;
|
|
use std::fs::FileType;
|
|
use tracing::Level;
|
|
|
|
#[cfg(target_family = "unix")]
|
|
use std::os::unix::ffi::OsStrExt;
|
|
|
|
use std::{
|
|
collections::HashMap,
|
|
fmt::Debug,
|
|
os::unix::prelude::PermissionsExt,
|
|
path::{Path, PathBuf},
|
|
};
|
|
use tokio_stream::StreamExt;
|
|
use tracing::instrument;
|
|
use walkdir::DirEntry;
|
|
use walkdir::WalkDir;
|
|
|
|
#[cfg(debug_assertions)]
|
|
use std::collections::HashSet;
|
|
|
|
#[derive(Debug, thiserror::Error)]
|
|
pub enum Error {
|
|
#[error("failed to upload directory at {0}: {1}")]
|
|
UploadDirectoryError(PathBuf, CastoreError),
|
|
|
|
#[error("invalid encoding encountered for entry {0:?}")]
|
|
InvalidEncoding(PathBuf),
|
|
|
|
#[error("unable to stat {0}: {1}")]
|
|
UnableToStat(PathBuf, std::io::Error),
|
|
|
|
#[error("unable to open {0}: {1}")]
|
|
UnableToOpen(PathBuf, std::io::Error),
|
|
|
|
#[error("unable to read {0}: {1}")]
|
|
UnableToRead(PathBuf, std::io::Error),
|
|
|
|
#[error("unsupported file {0} type: {1:?}")]
|
|
UnsupportedFileType(PathBuf, FileType),
|
|
}
|
|
|
|
impl From<CastoreError> for Error {
|
|
fn from(value: CastoreError) -> Self {
|
|
match value {
|
|
CastoreError::InvalidRequest(_) => panic!("tvix bug"),
|
|
CastoreError::StorageError(_) => panic!("error"),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl From<Error> for std::io::Error {
|
|
fn from(value: Error) -> Self {
|
|
std::io::Error::new(std::io::ErrorKind::Other, value)
|
|
}
|
|
}
|
|
|
|
/// This processes a given [walkdir::DirEntry] and returns a
|
|
/// proto::node::Node, depending on the type of the entry.
|
|
///
|
|
/// If the entry is a file, its contents are uploaded.
|
|
/// If the entry is a directory, the Directory is uploaded as well.
|
|
/// For this to work, it relies on the caller to provide the directory object
|
|
/// with the previously returned (child) nodes.
|
|
///
|
|
/// It assumes to be called only if all children of it have already been processed. If the entry is
|
|
/// indeed a directory, it'll also upload that directory to the store. For this, the
|
|
/// so-far-assembled Directory object for this path needs to be passed in.
|
|
///
|
|
/// It assumes the caller adds returned nodes to the directories it assembles.
|
|
#[instrument(skip_all, fields(entry.file_type=?&entry.file_type(),entry.path=?entry.path()))]
|
|
async fn process_entry<'a, BS>(
|
|
blob_service: BS,
|
|
directory_putter: &'a mut Box<dyn DirectoryPutter>,
|
|
entry: &'a walkdir::DirEntry,
|
|
maybe_directory: Option<Directory>,
|
|
) -> Result<Node, Error>
|
|
where
|
|
BS: AsRef<dyn BlobService> + Clone,
|
|
{
|
|
let file_type = entry.file_type();
|
|
|
|
if file_type.is_dir() {
|
|
let directory = maybe_directory
|
|
.expect("tvix bug: must be called with some directory in the case of directory");
|
|
let directory_digest = directory.digest();
|
|
let directory_size = directory.size();
|
|
|
|
// upload this directory
|
|
directory_putter
|
|
.put(directory)
|
|
.await
|
|
.map_err(|e| Error::UploadDirectoryError(entry.path().to_path_buf(), e))?;
|
|
|
|
return Ok(Node::Directory(DirectoryNode {
|
|
name: entry.file_name().as_bytes().to_owned().into(),
|
|
digest: directory_digest.into(),
|
|
size: directory_size,
|
|
}));
|
|
}
|
|
|
|
if file_type.is_symlink() {
|
|
let target: bytes::Bytes = std::fs::read_link(entry.path())
|
|
.map_err(|e| Error::UnableToStat(entry.path().to_path_buf(), e))?
|
|
.as_os_str()
|
|
.as_bytes()
|
|
.to_owned()
|
|
.into();
|
|
|
|
return Ok(Node::Symlink(SymlinkNode {
|
|
name: entry.file_name().as_bytes().to_owned().into(),
|
|
target,
|
|
}));
|
|
}
|
|
|
|
if file_type.is_file() {
|
|
let metadata = entry
|
|
.metadata()
|
|
.map_err(|e| Error::UnableToStat(entry.path().to_path_buf(), e.into()))?;
|
|
|
|
let mut file = tokio::fs::File::open(entry.path())
|
|
.await
|
|
.map_err(|e| Error::UnableToOpen(entry.path().to_path_buf(), e))?;
|
|
|
|
let mut writer = blob_service.as_ref().open_write().await;
|
|
|
|
if let Err(e) = tokio::io::copy(&mut file, &mut writer).await {
|
|
return Err(Error::UnableToRead(entry.path().to_path_buf(), e));
|
|
};
|
|
|
|
let digest = writer
|
|
.close()
|
|
.await
|
|
.map_err(|e| Error::UnableToRead(entry.path().to_path_buf(), e))?;
|
|
|
|
return Ok(Node::File(FileNode {
|
|
name: entry.file_name().as_bytes().to_vec().into(),
|
|
digest: digest.into(),
|
|
size: metadata.len(),
|
|
// If it's executable by the user, it'll become executable.
|
|
// This matches nix's dump() function behaviour.
|
|
executable: metadata.permissions().mode() & 64 != 0,
|
|
}));
|
|
}
|
|
|
|
// Nix says things like: error: file '/home/raito/dev/code.tvl.fyi/tvix/glue/src/tests/import_fixtures/a_devnode' has an unsupported type
|
|
Err(Error::UnsupportedFileType(
|
|
entry.path().to_path_buf(),
|
|
file_type,
|
|
))
|
|
}
|
|
|
|
/// Walk the filesystem at a given path and returns a level-keyed list of directory entries.
|
|
///
|
|
/// This is how [`ingest_path`] assembles the set of entries to pass on [`ingest_entries`].
|
|
/// This low-level function can be used if additional filtering or processing is required on the
|
|
/// entries.
|
|
///
|
|
/// Level here is in the context of graph theory, e.g. 2-level nodes
|
|
/// are nodes that are at depth 2.
|
|
///
|
|
/// This function will walk the filesystem using `walkdir` and will consume
|
|
/// `O(#number of entries)` space.
|
|
#[instrument(fields(path), err)]
|
|
pub fn walk_path_for_ingestion<P>(path: P) -> Result<Vec<Vec<DirEntry>>, Error>
|
|
where
|
|
P: AsRef<Path> + std::fmt::Debug,
|
|
{
|
|
let mut entries_per_depths: Vec<Vec<DirEntry>> = vec![Vec::new()];
|
|
for entry in WalkDir::new(path.as_ref())
|
|
.follow_links(false)
|
|
.follow_root_links(false)
|
|
.contents_first(false)
|
|
.sort_by_file_name()
|
|
.into_iter()
|
|
{
|
|
// Entry could be a NotFound, if the root path specified does not exist.
|
|
let entry = entry.map_err(|e| {
|
|
Error::UnableToOpen(
|
|
PathBuf::from(path.as_ref()),
|
|
e.into_io_error().expect("walkdir err must be some"),
|
|
)
|
|
})?;
|
|
|
|
if entry.depth() >= entries_per_depths.len() {
|
|
debug_assert!(
|
|
entry.depth() == entries_per_depths.len(),
|
|
"Received unexpected entry with depth {} during descent, previously at {}",
|
|
entry.depth(),
|
|
entries_per_depths.len()
|
|
);
|
|
|
|
entries_per_depths.push(vec![entry]);
|
|
} else {
|
|
entries_per_depths[entry.depth()].push(entry);
|
|
}
|
|
}
|
|
|
|
Ok(entries_per_depths)
|
|
}
|
|
|
|
/// Convert a leveled-key vector of filesystem entries into a stream of
|
|
/// [DirEntry] in a way that honors the Merkle invariant, i.e. from bottom to top.
|
|
pub fn leveled_entries_to_stream(
|
|
entries_per_depths: Vec<Vec<DirEntry>>,
|
|
) -> impl Stream<Item = DirEntry> {
|
|
stream! {
|
|
for level in entries_per_depths.into_iter().rev() {
|
|
for entry in level.into_iter() {
|
|
yield entry;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Ingests the contents at a given path into the tvix store, interacting with a [BlobService] and
|
|
/// [DirectoryService]. It returns the root node or an error.
|
|
///
|
|
/// It does not follow symlinks at the root, they will be ingested as actual symlinks.
|
|
#[instrument(skip(blob_service, directory_service), fields(path), err)]
|
|
pub async fn ingest_path<'a, BS, DS, P>(
|
|
blob_service: BS,
|
|
directory_service: DS,
|
|
path: P,
|
|
) -> Result<Node, Error>
|
|
where
|
|
P: AsRef<Path> + std::fmt::Debug,
|
|
BS: AsRef<dyn BlobService> + Clone,
|
|
DS: AsRef<dyn DirectoryService>,
|
|
{
|
|
let entries = walk_path_for_ingestion(path)?;
|
|
let entries_stream = leveled_entries_to_stream(entries);
|
|
pin_mut!(entries_stream);
|
|
|
|
ingest_entries(blob_service, directory_service, entries_stream).await
|
|
}
|
|
|
|
/// The Merkle invariant checker is an internal structure to perform bookkeeping of all directory
|
|
/// entries we are ingesting and verifying we are ingesting them in the right order.
|
|
///
|
|
/// That is, whenever we process an entry `L`, we would like to verify if we didn't process earlier
|
|
/// an entry `P` such that `P` is an **ancestor** of `L`.
|
|
///
|
|
/// If such a thing happened, it means that we have processed something like:
|
|
///
|
|
///```no_trust
|
|
/// A
|
|
/// / \
|
|
/// B C
|
|
/// / \ \
|
|
/// G F P <--------- processed before this one
|
|
/// / \ |
|
|
/// D E |
|
|
/// \ |
|
|
/// L <-----------------------------+
|
|
/// ```
|
|
///
|
|
/// This is exactly what must never happen.
|
|
///
|
|
/// Note: this checker is local, it can only see what happens on our side, not on the remote side,
|
|
/// i.e. the different remote services.
|
|
#[derive(Default)]
|
|
#[cfg(debug_assertions)]
|
|
struct MerkleInvariantChecker {
|
|
seen: HashSet<PathBuf>,
|
|
}
|
|
|
|
#[cfg(debug_assertions)]
|
|
impl MerkleInvariantChecker {
|
|
/// See a directory entry and remember it.
|
|
fn see(&mut self, node: &DirEntry) {
|
|
self.seen.insert(node.path().to_owned());
|
|
}
|
|
|
|
/// Returns a potential ancestor already seen for that directory entry.
|
|
fn find_ancestor(&self, node: &DirEntry) -> Option<PathBuf> {
|
|
for anc in node.path().ancestors() {
|
|
if self.seen.contains(anc) {
|
|
return Some(anc.to_owned());
|
|
}
|
|
}
|
|
|
|
None
|
|
}
|
|
}
|
|
|
|
/// Ingests elements from the given stream of [`DirEntry`] into a the passed [`BlobService`] and
|
|
/// [`DirectoryService`].
|
|
/// It does not follow symlinks at the root, they will be ingested as actual symlinks.
|
|
#[instrument(skip_all, ret(level = Level::TRACE), err)]
|
|
pub async fn ingest_entries<'a, BS, DS, S>(
|
|
blob_service: BS,
|
|
directory_service: DS,
|
|
mut entries_async_iterator: S,
|
|
) -> Result<Node, Error>
|
|
where
|
|
BS: AsRef<dyn BlobService> + Clone,
|
|
DS: AsRef<dyn DirectoryService>,
|
|
S: Stream<Item = DirEntry> + std::marker::Unpin,
|
|
{
|
|
let mut directories: HashMap<PathBuf, Directory> = HashMap::default();
|
|
|
|
let mut directory_putter = directory_service.as_ref().put_multiple_start();
|
|
|
|
#[cfg(debug_assertions)]
|
|
let mut invariant_checker: MerkleInvariantChecker = Default::default();
|
|
|
|
// We need to process a directory's children before processing
|
|
// the directory itself in order to have all the data needed
|
|
// to compute the hash.
|
|
while let Some(entry) = entries_async_iterator.next().await {
|
|
#[cfg(debug_assertions)]
|
|
{
|
|
// If we find an ancestor before we see this entry, this means that the caller
|
|
// broke the contract, refer to the documentation of the invariant checker to
|
|
// understand the reasoning here.
|
|
if let Some(ancestor) = invariant_checker.find_ancestor(&entry) {
|
|
panic!("Tvix bug: merkle invariant checker discovered that {} was processed before {}!",
|
|
ancestor.display(),
|
|
entry.path().display()
|
|
);
|
|
}
|
|
|
|
invariant_checker.see(&entry);
|
|
}
|
|
|
|
// FUTUREWORK: handle directory putting here, rather than piping it through process_entry.
|
|
let node = process_entry(
|
|
blob_service.clone(),
|
|
&mut directory_putter,
|
|
&entry,
|
|
// process_entry wants an Option<Directory> in case the entry points to a directory.
|
|
// make sure to provide it.
|
|
// If the directory has contents, we already have it in
|
|
// `directories` because we iterate over depth in reverse order (deepest to
|
|
// shallowest).
|
|
if entry.file_type().is_dir() {
|
|
Some(
|
|
directories
|
|
.remove(entry.path())
|
|
// In that case, it contained no children
|
|
.unwrap_or_default(),
|
|
)
|
|
} else {
|
|
None
|
|
},
|
|
)
|
|
.await?;
|
|
|
|
if entry.depth() == 0 {
|
|
// Make sure all the directories are flushed.
|
|
// FUTUREWORK: `debug_assert!` the resulting Ok(b3_digest) to be equal
|
|
// to `directories.get(entry.path())`.
|
|
if entry.file_type().is_dir() {
|
|
directory_putter.close().await?;
|
|
}
|
|
return Ok(node);
|
|
} else {
|
|
// calculate the parent path, and make sure we register the node there.
|
|
// NOTE: entry.depth() > 0
|
|
let parent_path = entry.path().parent().unwrap().to_path_buf();
|
|
|
|
// record node in parent directory, creating a new [proto:Directory] if not there yet.
|
|
let parent_directory = directories.entry(parent_path).or_default();
|
|
match node {
|
|
Node::Directory(e) => parent_directory.directories.push(e),
|
|
Node::File(e) => parent_directory.files.push(e),
|
|
Node::Symlink(e) => parent_directory.symlinks.push(e),
|
|
}
|
|
}
|
|
}
|
|
// unreachable, we already bailed out before if root doesn't exist.
|
|
unreachable!("Tvix bug: no root node emitted during ingestion")
|
|
}
|