|
|
|
@ -1,19 +1,26 @@
|
|
|
|
|
#pragma once
|
|
|
|
|
|
|
|
|
|
#include "arch/common.h"
|
|
|
|
|
#include "types.h"
|
|
|
|
|
|
|
|
|
|
#define ARCH_SIG_JMP_OFF 8
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* Initializes a context, zeros out registers, and sets the Instruction and
|
|
|
|
|
* Stack pointers
|
|
|
|
|
* Stack pointers. Sets variant to unused if ip and sp are 0, fast otherwise.
|
|
|
|
|
* @param actx arch_context to init
|
|
|
|
|
* @param ip value to set instruction pointer to
|
|
|
|
|
* @param sp value to set stack pointer to
|
|
|
|
|
*/
|
|
|
|
|
static void __attribute__((noinline)) arch_context_init(struct arch_context *actx, reg_t ip, reg_t sp)
|
|
|
|
|
{
|
|
|
|
|
assert(actx != NULL);
|
|
|
|
|
|
|
|
|
|
if (ip == 0 && sp == 0) {
|
|
|
|
|
actx->variant = arch_context_unused;
|
|
|
|
|
} else {
|
|
|
|
|
actx->variant = arch_context_fast;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (sp) {
|
|
|
|
|
/*
|
|
|
|
|
* context_switch conventions: bp is expected to be on top of the stack
|
|
|
|
@ -35,44 +42,73 @@ static void __attribute__((noinline)) arch_context_init(struct arch_context *act
|
|
|
|
|
|
|
|
|
|
actx->regs[ureg_rsp] = sp;
|
|
|
|
|
actx->regs[ureg_rip] = ip;
|
|
|
|
|
actx->variant = arch_context_fast;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* Restore a sandbox saved using a fastpath switch, restoring only the
|
|
|
|
|
* instruction pointer and stack pointer registers rather than
|
|
|
|
|
* a full mcontext, so it is less expensive than arch_mcontext_restore.
|
|
|
|
|
* @param active_context - the context of the current worker thread
|
|
|
|
|
* @param sandbox_context - the context that we want to restore
|
|
|
|
|
*/
|
|
|
|
|
static void
|
|
|
|
|
arch_context_restore(mcontext_t *active_context, struct arch_context *sandbox_context)
|
|
|
|
|
{
|
|
|
|
|
assert(active_context != NULL);
|
|
|
|
|
assert(sandbox_context != NULL);
|
|
|
|
|
|
|
|
|
|
/* Assumption: Base Context is only ever used by arch_context_switch */
|
|
|
|
|
assert(sandbox_context != &worker_thread_base_context);
|
|
|
|
|
|
|
|
|
|
assert(sandbox_context->regs[ureg_rsp]);
|
|
|
|
|
assert(sandbox_context->regs[ureg_rip]);
|
|
|
|
|
|
|
|
|
|
/* Transitioning from Fast -> Running */
|
|
|
|
|
assert(sandbox_context->variant == arch_context_fast);
|
|
|
|
|
sandbox_context->variant = arch_context_running;
|
|
|
|
|
|
|
|
|
|
active_context->gregs[REG_RSP] = sandbox_context->regs[ureg_rsp];
|
|
|
|
|
active_context->gregs[REG_RIP] = sandbox_context->regs[ureg_rip] + ARCH_SIG_JMP_OFF;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
|
* @param current - the registers and context of the thing running
|
|
|
|
|
* @param next - the registers and context of what we're switching to
|
|
|
|
|
* @param a - the registers and context of the thing running
|
|
|
|
|
* @param b - the registers and context of what we're switching to
|
|
|
|
|
* @return always returns 0, indicating success
|
|
|
|
|
*
|
|
|
|
|
* NULL in either of these values indicates the "no sandbox to execute" state,
|
|
|
|
|
* which defaults to resuming execution of main
|
|
|
|
|
*/
|
|
|
|
|
static inline int
|
|
|
|
|
arch_context_switch(struct arch_context *current, struct arch_context *next)
|
|
|
|
|
arch_context_switch(struct arch_context *a, struct arch_context *b)
|
|
|
|
|
{
|
|
|
|
|
/* Assumption: Software Interrupts are disabled by caller */
|
|
|
|
|
assert(software_interrupt_is_disabled);
|
|
|
|
|
|
|
|
|
|
if (next->variant == arch_context_fast && (next->regs[ureg_rip] == 0 || next->regs[ureg_rsp] == 0)) {
|
|
|
|
|
debuglog("Next Context was Fast Variant, but data was invalid.");
|
|
|
|
|
assert(0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/* if both current and next are NULL, there is no state change */
|
|
|
|
|
assert(current != NULL || next != NULL);
|
|
|
|
|
/* if both a and b are NULL, there is no state change */
|
|
|
|
|
assert(a != NULL || b != NULL);
|
|
|
|
|
|
|
|
|
|
/* Assumption: The caller does not switch to itself */
|
|
|
|
|
assert(current != next);
|
|
|
|
|
assert(a != b);
|
|
|
|
|
|
|
|
|
|
/* Set any NULLs to worker_thread_base_context to resume execution of main */
|
|
|
|
|
if (current == NULL) current = &worker_thread_base_context;
|
|
|
|
|
if (next == NULL) next = &worker_thread_base_context;
|
|
|
|
|
if (a == NULL) a = &worker_thread_base_context;
|
|
|
|
|
if (b == NULL) b = &worker_thread_base_context;
|
|
|
|
|
|
|
|
|
|
/* A Transition {Unused, Running} -> Fast */
|
|
|
|
|
assert(a->variant == arch_context_unused || a->variant == arch_context_running);
|
|
|
|
|
|
|
|
|
|
/* B Transition {Fast, Slow} -> Running */
|
|
|
|
|
assert(b->variant == arch_context_fast || b->variant == arch_context_slow);
|
|
|
|
|
|
|
|
|
|
/* Assumption: The context we are switching to should have saved a context in some form */
|
|
|
|
|
assert(next->variant == arch_context_fast || next->variant == arch_context_slow);
|
|
|
|
|
/* Assumption: Fastpath state is well formed */
|
|
|
|
|
if (b->variant == arch_context_fast) {
|
|
|
|
|
assert(b->regs[ureg_rip] != 0);
|
|
|
|
|
assert(b->regs[ureg_rsp] != 0);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
reg_t *current_registers = current->regs, *next_registers = next->regs;
|
|
|
|
|
assert(current_registers && next_registers);
|
|
|
|
|
reg_t *a_registers = a->regs, *b_registers = b->regs;
|
|
|
|
|
assert(a_registers && b_registers);
|
|
|
|
|
|
|
|
|
|
asm volatile(
|
|
|
|
|
/* Create a new stack frame */
|
|
|
|
@ -80,31 +116,40 @@ arch_context_switch(struct arch_context *current, struct arch_context *next)
|
|
|
|
|
"movq %%rsp, %%rbp\n\t" /* base_pointer = stack_pointer. Start new Frame */
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Save the IP and stack pointer to the context of the sandbox we're switching from
|
|
|
|
|
* Save Context A as a fastpath context switch
|
|
|
|
|
* Save the active stack pointer to context A.
|
|
|
|
|
* Set context A's IP to the absolute address of label 2f
|
|
|
|
|
*/
|
|
|
|
|
"movq $2f, 8(%%rax)\n\t" /* Write the address of label 2 to current_registers[1] (instruction_pointer). */
|
|
|
|
|
"movq %%rsp, (%%rax)\n\t" /* current_registers[0] (stack_pointer) = stack_pointer */
|
|
|
|
|
"movq $1, (%%rcx)\n\t" /* current->variant = arch_context_fast; */
|
|
|
|
|
"movq $2f, 8(%%rax)\n\t" /* Write the address of label 2 to context a's IP. */
|
|
|
|
|
"movq %%rsp, (%%rax)\n\t" /* a_registers[0] (stack_pointer) = stack_pointer */
|
|
|
|
|
"movq $1, (%%rcx)\n\t" /* a->variant = arch_context_fast; */
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Check if the variant of the context we're trying to switch to is SLOW (mcontext-based)
|
|
|
|
|
* If it is, jump to label 1 to restore the preempted sandbox
|
|
|
|
|
* Execute a fastpath or slowpath context switch based on context B's variant
|
|
|
|
|
*
|
|
|
|
|
* If slow (mcontext-based), jump to label 1 to restore via a signal handler
|
|
|
|
|
* Otherwise, fall through and execute fast path.
|
|
|
|
|
*/
|
|
|
|
|
"cmpq $2, (%%rdx)\n\t" /* if (next->variant == arch_context_slow); */
|
|
|
|
|
"cmpq $2, (%%rdx)\n\t" /* if (b->variant == arch_context_slow); */
|
|
|
|
|
"je 1f\n\t" /* goto 1; restore the existing sandbox using mcontext */
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Fast Path
|
|
|
|
|
* We can just write update the stack pointer and jump to the target instruction
|
|
|
|
|
*/
|
|
|
|
|
"movq (%%rbx), %%rsp\n\t" /* stack_pointer = next_registers[0] (stack_pointer) */
|
|
|
|
|
"jmpq *8(%%rbx)\n\t" /* immediate jump to next_registers[1] (instruction_pointer) */
|
|
|
|
|
"movq (%%rbx), %%rsp\n\t" /* stack_pointer = b_registers[0] (stack_pointer) */
|
|
|
|
|
"jmpq *8(%%rbx)\n\t" /* immediate jump to b_registers[1] (instruction_pointer) */
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* This is after the context switch. B is now the active context.
|
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Slow Path
|
|
|
|
|
* If the variant is arch_context_slow, that means the sandbox was preempted and we need to
|
|
|
|
|
* fallback to a full mcontext-based context switch. We do this by invoking
|
|
|
|
|
* arch_context_restore_preempted, which fires a SIGUSR1 signal. The SIGUSR1 signal handler
|
|
|
|
|
* arch_context_restore_preempted, which fires a SIGUSR1 signal. The SIGUSR1 signal handler
|
|
|
|
|
* executes the mcontext-based context switch.
|
|
|
|
|
*/
|
|
|
|
|
"1:\n\t"
|
|
|
|
@ -112,19 +157,20 @@ arch_context_switch(struct arch_context *current, struct arch_context *next)
|
|
|
|
|
".align 8\n\t"
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* This label was written to the instruction pointer of the sandbox that was switched away from
|
|
|
|
|
* These labels are used to switch
|
|
|
|
|
* This label is what is saved as the IP of a context that was saved using a fastpath context switch
|
|
|
|
|
* When this is resumed
|
|
|
|
|
* The sandbox either resumes at label 2 or 3 depending on if an offset of 8 is used.
|
|
|
|
|
*/
|
|
|
|
|
"2:\n\t"
|
|
|
|
|
"movq $3, (%%rdx)\n\t" /* next->variant = arch_context_running; */
|
|
|
|
|
"movq $3, (%%rdx)\n\t" /* b->variant = arch_context_running; */
|
|
|
|
|
".align 8\n\t"
|
|
|
|
|
|
|
|
|
|
/* This label is used in conjunction with a static offset */
|
|
|
|
|
"3:\n\t"
|
|
|
|
|
/* TODO: Should we set next->variant = arch_context_slow here?;*/
|
|
|
|
|
"popq %%rbp\n\t" /* base_pointer = stack[--stack_len]; Base Pointer is restored */
|
|
|
|
|
:
|
|
|
|
|
: "a"(current_registers), "b"(next_registers), "c"(¤t->variant), "d"(&next->variant)
|
|
|
|
|
: "a"(a_registers), "b"(b_registers), "c"(&a->variant), "d"(&b->variant)
|
|
|
|
|
: "memory", "cc", "rsi", "rdi", "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15", "xmm0", "xmm1", "xmm2",
|
|
|
|
|
"xmm3", "xmm4", "xmm5", "xmm6", "xmm7", "xmm8", "xmm9", "xmm10", "xmm11", "xmm12", "xmm13", "xmm14",
|
|
|
|
|
"xmm15");
|
|
|
|
|