This replaces the OpCode enum with a new Op enum which is guaranteed to fit in a
single byte. Instead of carrying enum variants with data, every variant that has
runtime data encodes it into the `Vec<u8>` that a `Chunk` now carries.
This has several advantages:
* Less stack space is required at runtime, and fewer allocations are required
while compiling.
* The OpCode doesn't need to carry "weird" special-cased data variants anymore.
* It is faster (albeit, not by much). On my laptop, results consistently look
approximately like this:
Benchmark 1: ./before -E '(import <nixpkgs> {}).firefox.outPath' --log-level ERROR --no-warnings
Time (mean ± σ): 8.224 s ± 0.272 s [User: 7.149 s, System: 0.688 s]
Range (min … max): 7.759 s … 8.583 s 10 runs
Benchmark 2: ./after -E '(import <nixpkgs> {}).firefox.outPath' --log-level ERROR --no-warnings
Time (mean ± σ): 8.000 s ± 0.198 s [User: 7.036 s, System: 0.633 s]
Range (min … max): 7.718 s … 8.334 s 10 runs
See notes below for why the performance impact might be less than expected.
* It is faster while at the same time dropping some optimisations we previously
performed.
This has several disadvantages:
* The code is closer to how one would write it in C or Go.
* Bit shifting!
* There is (for now) slightly more code than before.
On performance I have the following thoughts at the moment:
In order to prepare for adding GC, there's a couple of places in Tvix where I'd
like to fence off certain kinds of complexity (such as mutating bytecode, which,
for various reaons, also has to be part of data that is subject to GC). With
this change, we can drop optimisations like retroactively modifying existing
bytecode and *still* achieve better performance than before.
I believe that this is currently worth it to pave the way for changes that are
more significant for performance.
In general this also opens other avenues of optimisation: For example, we can
profile which argument sizes actually exist and remove the copy overhead of
varint decoding (which does show up in profiles) by using more adequately sized
types for, e.g., constant indices.
Known regressions:
* Op::Constant is no longer printing its values in disassembly (this can be
fixed, I just didn't get around to it, will do separately).
Change-Id: Id9b3a4254623a45de03069dbdb70b8349e976743
Reviewed-on: https://cl.tvl.fyi/c/depot/+/12191
Tested-by: BuildkiteCI
Reviewed-by: flokli <flokli@flokli.de>
427 lines
16 KiB
Rust
427 lines
16 KiB
Rust
//! This module implements the runtime representation of Thunks.
|
|
//!
|
|
//! Thunks are a special kind of Nix value, similar to a 0-argument
|
|
//! closure that yields some value. Thunks are used to implement the
|
|
//! lazy evaluation behaviour of Nix:
|
|
//!
|
|
//! Whenever the compiler determines that an expression should be
|
|
//! evaluated lazily, it creates a thunk instead of compiling the
|
|
//! expression value directly. At any point in the runtime where the
|
|
//! actual value of a thunk is required, it is "forced", meaning that
|
|
//! the encompassing computation takes place and the thunk takes on
|
|
//! its new value.
|
|
//!
|
|
//! Thunks have interior mutability to be able to memoise their
|
|
//! computation. Once a thunk is evaluated, its internal
|
|
//! representation becomes the result of the expression. It is legal
|
|
//! for the runtime to replace a thunk object directly with its value
|
|
//! object, but when forcing a thunk, the runtime *must* mutate the
|
|
//! memoisable slot.
|
|
|
|
use rustc_hash::FxHashSet;
|
|
use std::{
|
|
cell::{Ref, RefCell, RefMut},
|
|
fmt::Debug,
|
|
rc::Rc,
|
|
};
|
|
|
|
use crate::{
|
|
errors::ErrorKind,
|
|
opcode::Op,
|
|
upvalues::Upvalues,
|
|
value::Closure,
|
|
vm::generators::{self, GenCo},
|
|
Value,
|
|
};
|
|
|
|
use super::{Lambda, TotalDisplay};
|
|
use codemap::Span;
|
|
|
|
/// Internal representation of a suspended native thunk.
|
|
struct SuspendedNative(Box<dyn Fn() -> Result<Value, ErrorKind>>);
|
|
|
|
impl Debug for SuspendedNative {
|
|
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
|
write!(f, "SuspendedNative({:p})", self.0)
|
|
}
|
|
}
|
|
|
|
/// Internal representation of the different states of a thunk.
|
|
///
|
|
/// Upvalues must be finalised before leaving the initial state
|
|
/// (Suspended or RecursiveClosure). The [`value()`] function may
|
|
/// not be called until the thunk is in the final state (Evaluated).
|
|
#[derive(Debug)]
|
|
enum ThunkRepr {
|
|
/// Thunk is closed over some values, suspended and awaiting
|
|
/// execution.
|
|
Suspended {
|
|
lambda: Rc<Lambda>,
|
|
upvalues: Rc<Upvalues>,
|
|
span: Span,
|
|
},
|
|
|
|
/// Thunk is a suspended native computation.
|
|
Native(SuspendedNative),
|
|
|
|
/// Thunk currently under-evaluation; encountering a blackhole
|
|
/// value means that infinite recursion has occured.
|
|
Blackhole {
|
|
/// Span at which the thunk was first forced.
|
|
forced_at: Span,
|
|
|
|
/// Span at which the thunk was originally suspended.
|
|
suspended_at: Option<Span>,
|
|
|
|
/// Span of the first instruction of the actual code inside
|
|
/// the thunk.
|
|
content_span: Option<Span>,
|
|
},
|
|
|
|
// TODO(amjoseph): consider changing `Value` to `Rc<Value>` to avoid
|
|
// expensive clone()s in Thunk::force().
|
|
/// Fully evaluated thunk.
|
|
Evaluated(Value),
|
|
}
|
|
|
|
impl ThunkRepr {
|
|
fn debug_repr(&self) -> String {
|
|
match self {
|
|
ThunkRepr::Evaluated(v) => format!("thunk(val|{})", v),
|
|
ThunkRepr::Blackhole { .. } => "thunk(blackhole)".to_string(),
|
|
ThunkRepr::Native(_) => "thunk(native)".to_string(),
|
|
ThunkRepr::Suspended { lambda, .. } => format!("thunk({:p})", *lambda),
|
|
}
|
|
}
|
|
|
|
/// Return the Value within a fully-evaluated ThunkRepr; panics
|
|
/// if the thunk is not fully-evaluated.
|
|
fn expect(self) -> Value {
|
|
match self {
|
|
ThunkRepr::Evaluated(value) => value,
|
|
ThunkRepr::Blackhole { .. } => panic!("Thunk::expect() called on a black-holed thunk"),
|
|
ThunkRepr::Suspended { .. } | ThunkRepr::Native(_) => {
|
|
panic!("Thunk::expect() called on a suspended thunk")
|
|
}
|
|
}
|
|
}
|
|
|
|
fn expect_ref(&self) -> &Value {
|
|
match self {
|
|
ThunkRepr::Evaluated(value) => value,
|
|
ThunkRepr::Blackhole { .. } => panic!("Thunk::expect() called on a black-holed thunk"),
|
|
ThunkRepr::Suspended { .. } | ThunkRepr::Native(_) => {
|
|
panic!("Thunk::expect() called on a suspended thunk")
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn is_forced(&self) -> bool {
|
|
match self {
|
|
ThunkRepr::Evaluated(Value::Thunk(_)) => false,
|
|
ThunkRepr::Evaluated(_) => true,
|
|
_ => false,
|
|
}
|
|
}
|
|
}
|
|
|
|
/// A thunk is created for any value which requires non-strict
|
|
/// evaluation due to self-reference or lazy semantics (or both).
|
|
/// Every reference cycle involving `Value`s will contain at least
|
|
/// one `Thunk`.
|
|
#[derive(Clone, Debug)]
|
|
pub struct Thunk(Rc<RefCell<ThunkRepr>>);
|
|
|
|
impl Thunk {
|
|
pub fn new_closure(lambda: Rc<Lambda>) -> Self {
|
|
Thunk(Rc::new(RefCell::new(ThunkRepr::Evaluated(Value::Closure(
|
|
Rc::new(Closure {
|
|
upvalues: Rc::new(Upvalues::with_capacity(lambda.upvalue_count)),
|
|
lambda: lambda.clone(),
|
|
}),
|
|
)))))
|
|
}
|
|
|
|
pub fn new_suspended(lambda: Rc<Lambda>, span: Span) -> Self {
|
|
Thunk(Rc::new(RefCell::new(ThunkRepr::Suspended {
|
|
upvalues: Rc::new(Upvalues::with_capacity(lambda.upvalue_count)),
|
|
lambda: lambda.clone(),
|
|
span,
|
|
})))
|
|
}
|
|
|
|
pub fn new_suspended_native(native: Box<dyn Fn() -> Result<Value, ErrorKind>>) -> Self {
|
|
Thunk(Rc::new(RefCell::new(ThunkRepr::Native(SuspendedNative(
|
|
native,
|
|
)))))
|
|
}
|
|
|
|
/// Helper function to create a [`Thunk`] that calls a function given as the
|
|
/// [`Value`] `callee` with the argument `arg` when it is forced. This is
|
|
/// particularly useful in builtin implementations if the result of calling
|
|
/// a function does not need to be forced immediately, because e.g. it is
|
|
/// stored in an attribute set.
|
|
pub fn new_suspended_call(callee: Value, arg: Value, span: Span) -> Self {
|
|
let mut lambda = Lambda::default();
|
|
|
|
let arg_idx = lambda.chunk().push_constant(arg);
|
|
let f_idx = lambda.chunk().push_constant(callee);
|
|
|
|
// This is basically a recreation of compile_apply():
|
|
// We need to push the argument onto the stack and then the function.
|
|
// The function (not the argument) needs to be forced before calling.
|
|
lambda.chunk.push_op(Op::Constant, span);
|
|
lambda.chunk.push_uvarint(arg_idx.0 as u64);
|
|
lambda.chunk.push_op(Op::Constant, span);
|
|
lambda.chunk.push_uvarint(f_idx.0 as u64);
|
|
lambda.chunk.push_op(Op::Force, span);
|
|
lambda.chunk.push_op(Op::Call, span);
|
|
|
|
// Inform the VM that the chunk has ended
|
|
lambda.chunk.push_op(Op::Return, span);
|
|
|
|
Thunk(Rc::new(RefCell::new(ThunkRepr::Suspended {
|
|
upvalues: Rc::new(Upvalues::with_capacity(0)),
|
|
lambda: Rc::new(lambda),
|
|
span,
|
|
})))
|
|
}
|
|
|
|
fn prepare_blackhole(&self, forced_at: Span) -> ThunkRepr {
|
|
match &*self.0.borrow() {
|
|
ThunkRepr::Suspended { span, lambda, .. } => ThunkRepr::Blackhole {
|
|
forced_at,
|
|
suspended_at: Some(*span),
|
|
content_span: Some(lambda.chunk.first_span()),
|
|
},
|
|
|
|
_ => ThunkRepr::Blackhole {
|
|
forced_at,
|
|
suspended_at: None,
|
|
content_span: None,
|
|
},
|
|
}
|
|
}
|
|
|
|
pub async fn force(myself: Thunk, co: GenCo, span: Span) -> Result<Value, ErrorKind> {
|
|
Self::force_(myself, &co, span).await
|
|
}
|
|
pub async fn force_(mut myself: Thunk, co: &GenCo, span: Span) -> Result<Value, ErrorKind> {
|
|
// This vector of "thunks which point to the thunk-being-forced", to
|
|
// be updated along with it, is necessary in order to write this
|
|
// function in iterative (and later, mutual-tail-call) form.
|
|
let mut also_update: Vec<Rc<RefCell<ThunkRepr>>> = vec![];
|
|
|
|
loop {
|
|
// If the current thunk is already fully evaluated, return its evaluated
|
|
// value. The VM will continue running the code that landed us here.
|
|
if myself.is_forced() {
|
|
let val = myself.unwrap_or_clone();
|
|
for other_thunk in also_update.into_iter() {
|
|
other_thunk.replace(ThunkRepr::Evaluated(val.clone()));
|
|
}
|
|
return Ok(val);
|
|
}
|
|
|
|
// Begin evaluation of this thunk by marking it as a blackhole, meaning
|
|
// that any other forcing frame encountering this thunk before its
|
|
// evaluation is completed detected an evaluation cycle.
|
|
let inner = myself.0.replace(myself.prepare_blackhole(span));
|
|
|
|
match inner {
|
|
// If there was already a blackhole in the thunk, this is an
|
|
// evaluation cycle.
|
|
ThunkRepr::Blackhole {
|
|
forced_at,
|
|
suspended_at,
|
|
content_span,
|
|
} => {
|
|
return Err(ErrorKind::InfiniteRecursion {
|
|
first_force: forced_at,
|
|
suspended_at,
|
|
content_span,
|
|
})
|
|
}
|
|
|
|
// If there is a native function stored in the thunk, evaluate it
|
|
// and replace this thunk's representation with the result.
|
|
ThunkRepr::Native(native) => {
|
|
let value = native.0()?;
|
|
myself.0.replace(ThunkRepr::Evaluated(value));
|
|
continue;
|
|
}
|
|
|
|
// When encountering a suspended thunk, request that the VM enters
|
|
// it and produces the result.
|
|
ThunkRepr::Suspended {
|
|
lambda,
|
|
upvalues,
|
|
span,
|
|
} => {
|
|
// TODO(amjoseph): use #[tailcall::mutual] here. This can
|
|
// be turned into a tailcall to vm::execute_bytecode() by
|
|
// passing `also_update` to it.
|
|
let value = generators::request_enter_lambda(co, lambda, upvalues, span).await;
|
|
myself.0.replace(ThunkRepr::Evaluated(value));
|
|
continue;
|
|
}
|
|
|
|
// nested thunks -- try to flatten before forcing
|
|
ThunkRepr::Evaluated(Value::Thunk(inner_thunk)) => {
|
|
match Rc::try_unwrap(inner_thunk.0) {
|
|
Ok(refcell) => {
|
|
// we are the only reference to the inner thunk,
|
|
// so steal it
|
|
myself.0.replace(refcell.into_inner());
|
|
continue;
|
|
}
|
|
Err(rc) => {
|
|
let inner_thunk = Thunk(rc);
|
|
if inner_thunk.is_forced() {
|
|
// tail call to force the inner thunk; note that
|
|
// this means the outer thunk remains unforced
|
|
// even after calling force() on it; however the
|
|
// next time it is forced we will be one
|
|
// thunk-forcing closer to it being
|
|
// fully-evaluated.
|
|
myself
|
|
.0
|
|
.replace(ThunkRepr::Evaluated(inner_thunk.value().clone()));
|
|
continue;
|
|
}
|
|
also_update.push(myself.0.clone());
|
|
myself = inner_thunk;
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
ThunkRepr::Evaluated(val) => {
|
|
return Ok(val);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn finalise(&self, stack: &[Value]) {
|
|
self.upvalues_mut().resolve_deferred_upvalues(stack);
|
|
}
|
|
|
|
pub fn is_evaluated(&self) -> bool {
|
|
matches!(*self.0.borrow(), ThunkRepr::Evaluated(_))
|
|
}
|
|
|
|
pub fn is_suspended(&self) -> bool {
|
|
matches!(
|
|
*self.0.borrow(),
|
|
ThunkRepr::Suspended { .. } | ThunkRepr::Native(_)
|
|
)
|
|
}
|
|
|
|
/// Returns true if forcing this thunk will not change it.
|
|
pub fn is_forced(&self) -> bool {
|
|
self.0.borrow().is_forced()
|
|
}
|
|
|
|
/// Returns a reference to the inner evaluated value of a thunk.
|
|
/// It is an error to call this on a thunk that has not been
|
|
/// forced, or is not otherwise known to be fully evaluated.
|
|
// Note: Due to the interior mutability of thunks this is
|
|
// difficult to represent in the type system without impacting the
|
|
// API too much.
|
|
pub fn value(&self) -> Ref<Value> {
|
|
Ref::map(self.0.borrow(), |thunk| match thunk {
|
|
ThunkRepr::Evaluated(value) => value,
|
|
ThunkRepr::Blackhole { .. } => panic!("Thunk::value called on a black-holed thunk"),
|
|
ThunkRepr::Suspended { .. } | ThunkRepr::Native(_) => {
|
|
panic!("Thunk::value called on a suspended thunk")
|
|
}
|
|
})
|
|
}
|
|
|
|
/// Returns the inner evaluated value of a thunk, cloning it if
|
|
/// the Rc has more than one strong reference. It is an error
|
|
/// to call this on a thunk that has not been forced, or is not
|
|
/// otherwise known to be fully evaluated.
|
|
fn unwrap_or_clone(self) -> Value {
|
|
match Rc::try_unwrap(self.0) {
|
|
Ok(refcell) => refcell.into_inner().expect(),
|
|
Err(rc) => Ref::map(rc.borrow(), |thunkrepr| thunkrepr.expect_ref()).clone(),
|
|
}
|
|
}
|
|
|
|
pub fn upvalues(&self) -> Ref<'_, Upvalues> {
|
|
Ref::map(self.0.borrow(), |thunk| match thunk {
|
|
ThunkRepr::Suspended { upvalues, .. } => upvalues.as_ref(),
|
|
ThunkRepr::Evaluated(Value::Closure(c)) => &c.upvalues,
|
|
_ => panic!("upvalues() on non-suspended thunk"),
|
|
})
|
|
}
|
|
|
|
pub fn upvalues_mut(&self) -> RefMut<'_, Upvalues> {
|
|
RefMut::map(self.0.borrow_mut(), |thunk| match thunk {
|
|
ThunkRepr::Suspended { upvalues, .. } => Rc::get_mut(upvalues).unwrap(),
|
|
ThunkRepr::Evaluated(Value::Closure(c)) => Rc::get_mut(
|
|
&mut Rc::get_mut(c).unwrap().upvalues,
|
|
)
|
|
.expect(
|
|
"upvalues_mut() was called on a thunk which already had multiple references to it",
|
|
),
|
|
thunk => panic!("upvalues() on non-suspended thunk: {thunk:?}"),
|
|
})
|
|
}
|
|
|
|
/// Do not use this without first reading and understanding
|
|
/// `tvix/docs/value-pointer-equality.md`.
|
|
pub(crate) fn ptr_eq(&self, other: &Self) -> bool {
|
|
if Rc::ptr_eq(&self.0, &other.0) {
|
|
return true;
|
|
}
|
|
match &*self.0.borrow() {
|
|
ThunkRepr::Evaluated(Value::Closure(c1)) => match &*other.0.borrow() {
|
|
ThunkRepr::Evaluated(Value::Closure(c2)) => Rc::ptr_eq(c1, c2),
|
|
_ => false,
|
|
},
|
|
_ => false,
|
|
}
|
|
}
|
|
|
|
/// Helper function to format thunks in observer output.
|
|
pub(crate) fn debug_repr(&self) -> String {
|
|
self.0.borrow().debug_repr()
|
|
}
|
|
}
|
|
|
|
impl TotalDisplay for Thunk {
|
|
fn total_fmt(&self, f: &mut std::fmt::Formatter<'_>, set: &mut ThunkSet) -> std::fmt::Result {
|
|
if !set.insert(self) {
|
|
return f.write_str("<CYCLE>");
|
|
}
|
|
|
|
match &*self.0.borrow() {
|
|
ThunkRepr::Evaluated(v) => v.total_fmt(f, set),
|
|
ThunkRepr::Suspended { .. } | ThunkRepr::Native(_) => f.write_str("<CODE>"),
|
|
other => write!(f, "internal[{}]", other.debug_repr()),
|
|
}
|
|
}
|
|
}
|
|
|
|
/// A wrapper type for tracking which thunks have already been seen
|
|
/// in a context. This is necessary for printing and deeply forcing
|
|
/// cyclic non-diverging data structures like `rec { f = [ f ]; }`.
|
|
/// This is separate from the ThunkRepr::Blackhole mechanism, which
|
|
/// detects diverging data structures like `(rec { f = f; }).f`.
|
|
///
|
|
/// The inner `HashSet` is not available on the outside, as it would be
|
|
/// potentially unsafe to interact with the pointers in the set.
|
|
#[derive(Default)]
|
|
pub struct ThunkSet(FxHashSet<*const ThunkRepr>);
|
|
|
|
impl ThunkSet {
|
|
/// Check whether the given thunk has already been seen. Will mark the thunk
|
|
/// as seen otherwise.
|
|
pub fn insert(&mut self, thunk: &Thunk) -> bool {
|
|
let ptr: *const ThunkRepr = thunk.0.as_ptr();
|
|
self.0.insert(ptr)
|
|
}
|
|
}
|