1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
use crate::error::OutOfMemory;
use crate::runtime::vm::vmcontext::VMArrayCallNative;
use crate::runtime::vm::{
StoreBox, TrapRegisters, TrapTest, VMContext, VMOpaqueContext, f32x4, f64x2, i8x16, tls,
};
use crate::{Engine, ValRaw};
use core::marker;
use core::ptr::NonNull;
use pulley_interpreter::interp::{DoneReason, RegType, TrapKind, Val, Vm, XRegVal};
use pulley_interpreter::{Reg, XReg};
use wasmtime_environ::{BuiltinFunctionIndex, HostCall, Trap};
use wasmtime_unwinder::Handler;
use wasmtime_unwinder::Unwind;
/// Interpreter state stored within a `Store<T>`.
#[repr(transparent)]
pub struct Interpreter {
/// Pulley VM state, stored in a `StoreBox<T>`.
///
/// This representation has a dual purpose of (a) having a low overhead if
/// pulley is disabled (just a null pointer) and (b) enabling safe access to
/// the `Vm` in the face of recursive calls.
///
/// For (b) that's the most tricky part of this, but the basic problem looks
/// like:
///
/// * The host initially executes some WebAssembly.
/// * This acquires a `&mut Vm` and does some execution.
/// * The WebAssembly then invokes the host.
/// * This bottoms out in `CallIndirectHost` which means that we'll do a
/// dynamic dispatch to a function pointer in pulley registers.
/// * The function we call gets unfettered access to `StoreContextMut<T>`
/// * When the function returns our original `&mut Vm` pointer is
/// invalidated, so it has to be re-acquired.
///
/// The usage of `StoreBox` here solves this conundrum by storing the
/// `InterpreterRef` at-rest as a `NonNull<Vm>` as opposed to a `&mut Vm`.
/// This is required to model how after a host call the VM state must be
/// re-acquire from store state to re-assert that it has an exclusive
/// borrow.
///
/// This in turn models how VM state could be modified as part of the
/// recursive function call, for example with another VM execution itself.
///
/// Note that the safety of this all relies not only on correctly managing
/// this pointer but it also requires that this pointer is never
/// deallocated while an `InterpreterRef` is live. The `InterpreterRef` type
/// carries a borrow of this type to ensure this isn't dropped
/// independently, and then this file never overwrites this private field to
/// otherwise guarantee this.
pulley: StoreBox<VmState>,
}
struct VmState {
vm: Vm,
resume_at_pc: Option<usize>,
}
impl Interpreter {
/// Creates a new interpreter ready to interpret code.
pub fn new(engine: &Engine) -> Result<Interpreter, OutOfMemory> {
let ret = Interpreter {
pulley: StoreBox::new(VmState {
vm: Vm::with_stack(engine.config().max_wasm_stack)?,
resume_at_pc: None,
})?,
};
engine.profiler().register_interpreter(&ret);
Ok(ret)
}
/// Returns the `InterpreterRef` structure which can be used to actually
/// execute interpreted code.
pub fn as_interpreter_ref(&mut self) -> InterpreterRef<'_> {
InterpreterRef {
vm: self.pulley.get(),
_phantom: marker::PhantomData,
}
}
pub fn pulley(&self) -> &Vm {
let state = unsafe { self.pulley.get().as_ref() };
&state.vm
}
/// Get an implementation of `Unwind` used to walk the Pulley stack.
pub fn unwinder(&self) -> &'static dyn Unwind {
&UnwindPulley
}
}
/// Wrapper around `&mut pulley_interpreter::Vm` to enable compiling this to a
/// zero-sized structure when pulley is disabled at compile time.
#[repr(transparent)]
pub struct InterpreterRef<'a> {
vm: NonNull<VmState>,
_phantom: marker::PhantomData<&'a mut VmState>,
}
/// An implementation of stack-walking details specifically designed
/// for unwinding Pulley's runtime stack.
pub struct UnwindPulley;
unsafe impl Unwind for UnwindPulley {
fn next_older_fp_from_fp_offset(&self) -> usize {
0
}
fn next_older_sp_from_fp_offset(&self) -> usize {
if cfg!(target_pointer_width = "32") {
8
} else {
16
}
}
unsafe fn get_next_older_pc_from_fp(&self, fp: usize) -> usize {
// The calling convention always pushes the return pointer (aka the PC
// of the next older frame) just before this frame.
unsafe { *(fp as *mut usize).offset(1) }
}
fn assert_fp_is_aligned(&self, fp: usize) {
let expected = if cfg!(target_pointer_width = "32") {
8
} else {
16
};
assert_eq!(fp % expected, 0, "stack should always be aligned");
}
}
impl InterpreterRef<'_> {
fn vm_state(&mut self) -> &mut VmState {
// SAFETY: This is a bit of a tricky code. The safety here is isolated
// to this file, but not isolated to just this function call.
//
// An `InterpreterRef` guarantees that we have a pointer to a `Vm`, and
// that pointer originates from a `StoreBox<VM>` in the store itself.
// One level of safety here relies on that never being deallocated or
// overwritten, which this file upholds as it's a private field only
// this module can access.
//
// Another aspect upheld by `InterpreterRef` is that it transfers, to
// the compiler, a mutable borrow of the store (e.g `struct Interpreter`
// above) to this reference. While this doesn't actually hold such a
// lifetime-bound pointer it guarantees that only one of these can be
// active at a time per interpreter.
//
// Finally the lifetime of the returned `Vm` is bound to `self` which
// ensures that there is at most one per `InterpreterRef`.
//
// All put together this should allow at most one `&mut Vm` per-store,
// which is one guarantee we need for this to be safe.
//
// Otherwise this is then done to represent how across host function
// calls the interpreter needs to be re-borrowed as the state may have
// changed as part of the dynamic host call.
unsafe { self.vm.as_mut() }
}
fn vm(&mut self) -> &mut Vm {
&mut self.vm_state().vm
}
/// Invokes interpreted code.
///
/// The `bytecode` pointer should previously have been produced by Cranelift
/// and `callee` / `caller` / `args_and_results` are normal array-call
/// arguments being passed around.
pub unsafe fn call(
mut self,
mut bytecode: NonNull<u8>,
callee: NonNull<VMOpaqueContext>,
caller: NonNull<VMContext>,
args_and_results: NonNull<[ValRaw]>,
) -> bool {
// Initialize argument registers with the ABI arguments.
let args = [
XRegVal::new_ptr(callee.as_ptr()).into(),
XRegVal::new_ptr(caller.as_ptr()).into(),
XRegVal::new_ptr(args_and_results.cast::<u8>().as_ptr()).into(),
XRegVal::new_u64(args_and_results.len() as u64).into(),
];
let mut vm = self.vm();
let old_lr = unsafe { vm.call_start(&args) };
// Run the interpreter as much as possible until it finishes, and then
// handle each finish condition differently.
let ret = loop {
match unsafe { vm.call_run(bytecode) } {
// If the VM returned entirely then read the return value and
// return that (it indicates whether a trap happened or not.
DoneReason::ReturnToHost(()) => {
match unsafe { vm.call_end(old_lr, [RegType::XReg]).next().unwrap() } {
#[allow(
clippy::cast_possible_truncation,
reason = "intentionally reading the lower bits only"
)]
Val::XReg(xreg) => break (xreg.get_u32() as u8) != 0,
_ => unreachable!(),
}
}
// If the VM wants to call out to the host then dispatch that
// here based on `id`. Once that returns we typically resume
// execution at `resume`.
DoneReason::CallIndirectHost { id, resume } => {
unsafe {
self.call_indirect_host(id);
}
// After the host has finished take a look at what hostcall
// was just made. The `raise` hostcall gets special handling
// for its non-local transfer of control flow.
//
// Also note that for non-`raise` hostcalls the
// `state.resume_at_pc` value should always be `None`.
if u32::from(id) == HostCall::Builtin(BuiltinFunctionIndex::raise()).index() {
bytecode = self.take_resume_at_pc();
} else {
debug_assert!(self.vm_state().resume_at_pc.is_none());
bytecode = resume;
}
vm = self.vm();
}
// If the VM trapped then process that here and return `false`.
DoneReason::Trap { pc, kind } => {
bytecode = self.trap(pc, kind);
vm = self.vm();
}
}
};
ret
}
/// Handles the `call_indirect_host` instruction, dispatching the `sig`
/// number here which corresponds to `wasmtime_environ::HostCall`.
#[allow(
clippy::cast_possible_truncation,
clippy::cast_sign_loss,
unused,
reason = "macro-generated code"
)]
#[cfg_attr(
not(feature = "component-model"),
expect(unused_macro_rules, reason = "macro-code")
)]
unsafe fn call_indirect_host(&mut self, id: u8) {
let id = u32::from(id);
let fnptr = self.vm()[XReg::x0].get_ptr();
let mut arg_reg = 1;
/// Helper macro to invoke a builtin.
///
/// Used as:
///
/// `call(@builtin(ty1, ty2, ...) -> retty)` - invoke a core or
/// component builtin with the macro-defined signature.
///
/// `call(@host Ty(ty1, ty2, ...) -> retty)` - invoke a host function
/// with the type `Ty`. The other types in the macro are checked by
/// rustc to match the actual `Ty` definition in Rust.
macro_rules! call {
(@builtin($($param:ident),*) $(-> $result:ident)?) => {{
#[allow(improper_ctypes_definitions, reason = "__m128i known not FFI-safe")]
type T = unsafe extern "C" fn($(call!(@ty $param)),*) $(-> call!(@ty $result))?;
call!(@host T($($param),*) $(-> $result)?);
}};
(@host $ty:ident($($param:ident),*) $(-> $result:ident)?) => {{
// Decode each argument according to this macro, pulling
// arguments from successive registers.
let ret = unsafe {
let mut vm = self.vm();
// Convert the pointer from pulley to a native function pointer.
union GetNative {
fnptr: *mut u8,
host: $ty,
}
let host = GetNative { fnptr }.host;
host($({
let reg = XReg::new(arg_reg).unwrap();
arg_reg += 1;
call!(@get $param vm[reg])
}),*)
};
let _ = arg_reg; // silence last dead arg_reg increment warning
let state = self.vm_state();
let _vm = &mut state.vm;
// Store the return value, if one is here, in x0.
$(
call!(@set $result ret => _vm[XReg::x0]);
)?
let _ = ret; // silence warning if no return value
// Return from the outer `call_indirect_host` host function as
// it's been processed.
return;
}};
// Conversion from macro-defined types to Rust host types.
(@ty bool) => (bool);
(@ty u8) => (u8);
(@ty u32) => (u32);
(@ty i32) => (i32);
(@ty u64) => (u64);
(@ty i64) => (i64);
(@ty f32) => (f32);
(@ty f64) => (f64);
(@ty i8x16) => (i8x16);
(@ty f32x4) => (f32x4);
(@ty f64x2) => (f64x2);
(@ty vmctx) => (*mut VMContext);
(@ty pointer) => (*mut u8);
(@ty ptr_u8) => (*mut u8);
(@ty ptr_u16) => (*mut u16);
(@ty ptr_size) => (*mut usize);
(@ty size) => (usize);
// Conversion from a pulley register value to the macro-defined
// type.
(@get u8 $reg:expr) => ($reg.get_i32() as u8);
(@get u32 $reg:expr) => ($reg.get_u32());
(@get u64 $reg:expr) => ($reg.get_u64());
(@get f32 $reg:expr) => (unreachable::<f32, _>($reg));
(@get f64 $reg:expr) => (unreachable::<f64, _>($reg));
(@get i8x16 $reg:expr) => (unreachable::<i8x16, _>($reg));
(@get f32x4 $reg:expr) => (unreachable::<f32x4, _>($reg));
(@get f64x2 $reg:expr) => (unreachable::<f64x2, _>($reg));
(@get vmctx $reg:expr) => ($reg.get_ptr());
(@get pointer $reg:expr) => ($reg.get_ptr());
(@get ptr $reg:expr) => ($reg.get_ptr());
(@get nonnull $reg:expr) => (NonNull::new($reg.get_ptr()).unwrap());
(@get ptr_u8 $reg:expr) => ($reg.get_ptr());
(@get ptr_u16 $reg:expr) => ($reg.get_ptr());
(@get ptr_size $reg:expr) => ($reg.get_ptr());
(@get size $reg:expr) => ($reg.get_ptr::<u8>() as usize);
// Conversion from a Rust value back into a macro-defined type,
// stored in a pulley register.
(@set bool $src:expr => $dst:expr) => ($dst.set_i32(i32::from($src)));
(@set u32 $src:expr => $dst:expr) => ($dst.set_u32($src));
(@set u64 $src:expr => $dst:expr) => ($dst.set_u64($src));
(@set f32 $src:expr => $dst:expr) => (unreachable::<f32, _>(($dst, $src)));
(@set f64 $src:expr => $dst:expr) => (unreachable::<f64, _>(($dst, $src)));
(@set i8x16 $src:expr => $dst:expr) => (unreachable::<i8x16, _>(($dst, $src)));
(@set f32x4 $src:expr => $dst:expr) => (unreachable::<f32x4, _>(($dst, $src)));
(@set f64x2 $src:expr => $dst:expr) => (unreachable::<f64x2, _>(($dst, $src)));
(@set pointer $src:expr => $dst:expr) => ($dst.set_ptr($src));
(@set size $src:expr => $dst:expr) => ($dst.set_ptr($src as *mut u8));
}
// With the helper macro above structure this into:
//
// foreach [core, component]
// * dispatch the call-the-host function pointer type
// * dispatch all builtins by their index.
//
// The hope is that this is relatively easy for LLVM to optimize since
// it's a bunch of:
//
// if id == 0 { ...; return; }
// if id == 1 { ...; return; }
// if id == 2 { ...; return; }
// ...
//
if id == const { HostCall::ArrayCall.index() } {
call!(@host VMArrayCallNative(nonnull, nonnull, nonnull, size) -> bool);
}
macro_rules! core {
(
$(
$( #[cfg($attr:meta)] )?
$name:ident($($pname:ident: $param:ident ),* ) $(-> $result:ident)?;
)*
) => {
$(
$( #[cfg($attr)] )?
if id == const { HostCall::Builtin(BuiltinFunctionIndex::$name()).index() } {
call!(@builtin($($param),*) $(-> $result)?);
}
)*
}
}
wasmtime_environ::foreach_builtin_function!(core);
#[cfg(feature = "component-model")]
{
use crate::runtime::vm::component::VMLoweringCallee;
use wasmtime_environ::component::ComponentBuiltinFunctionIndex;
if id == const { HostCall::ComponentLowerImport.index() } {
call!(@host VMLoweringCallee(nonnull, nonnull, u32, u32, nonnull, size) -> bool);
}
macro_rules! component {
(
$(
$( #[cfg($attr:meta)] )?
$name:ident($($pname:ident: $param:ident ),* ) $(-> $result:ident)?;
)*
) => {
$(
$( #[cfg($attr)] )?
if id == const { HostCall::ComponentBuiltin(ComponentBuiltinFunctionIndex::$name()).index() } {
call!(@builtin($($param),*) $(-> $result)?);
}
)*
}
}
wasmtime_environ::foreach_builtin_component_function!(component);
}
// if we got this far then something has gone seriously wrong.
return unreachable(());
fn unreachable<T, U>(_: U) -> T {
unreachable!()
}
}
/// Configures Pulley to be able to resume to the specified exception
/// handler.
///
/// This is executed from a `raise` hostcall when an exception is being
/// raised.
///
/// # Safety
///
/// Requires that all the parameters here are valid and will leave Pulley
/// in a valid state for executing.
pub(crate) unsafe fn resume_to_exception_handler(
&mut self,
handler: &Handler,
payload1: usize,
payload2: usize,
) {
unsafe {
let vm = self.vm();
vm[XReg::x0].set_u64(payload1 as u64);
vm[XReg::x1].set_u64(payload2 as u64);
vm[XReg::sp].set_ptr(core::ptr::with_exposed_provenance_mut::<u8>(handler.sp));
vm.set_fp(core::ptr::with_exposed_provenance_mut(handler.fp));
}
let state = self.vm_state();
debug_assert!(state.resume_at_pc.is_none());
self.vm_state().resume_at_pc = Some(handler.pc);
}
/// Handles an interpreter trap. This will initialize the trap state stored
/// in TLS via the `test_if_trap` helper below by reading the pc/fp of the
/// interpreter and seeing if that's a valid opcode to trap at.
fn trap(&mut self, pc: NonNull<u8>, kind: Option<TrapKind>) -> NonNull<u8> {
let regs = TrapRegisters {
pc: pc.as_ptr() as usize,
fp: self.vm().fp() as usize,
};
let handler = tls::with(|s| {
let s = s.unwrap();
match kind {
Some(kind) => {
let trap = match kind {
TrapKind::IntegerOverflow => Trap::IntegerOverflow,
TrapKind::DivideByZero => Trap::IntegerDivisionByZero,
TrapKind::BadConversionToInteger => Trap::BadConversionToInteger,
TrapKind::MemoryOutOfBounds => Trap::MemoryOutOfBounds,
TrapKind::DisabledOpcode => Trap::DisabledOpcode,
TrapKind::StackOverflow => Trap::StackOverflow,
};
s.set_jit_trap(regs, None, trap);
s.entry_trap_handler()
}
None => {
match s.test_if_trap(regs, None, |_| false) {
// This shouldn't be possible, so this is a fatal error
// if it happens.
TrapTest::NotWasm => {
panic!("pulley trap at {pc:?} without trap code registered")
}
// Not possible with our closure above returning `false`.
#[cfg(has_host_compiler_backend)]
TrapTest::HandledByEmbedder => unreachable!(),
// Trap was handled, yay! Configure interpreter state
// to resume at the exception handler.
TrapTest::Trap(handler) => handler,
}
}
}
});
unsafe {
self.resume_to_exception_handler(&handler, 0, 0);
}
self.take_resume_at_pc()
}
fn take_resume_at_pc(&mut self) -> NonNull<u8> {
let pc = self.vm_state().resume_at_pc.take().unwrap();
let pc = core::ptr::with_exposed_provenance_mut(pc);
NonNull::new(pc).unwrap()
}
}