#pragma once #include #include #include #include #include #include #include "software_interrupt.h" #define ARCH_SIG_JMP_OFF 8 // Userspace Registers. enum UREGS { UREG_RSP = 0, UREG_RIP = 1, UREG_COUNT }; typedef uint64_t reg_t; /* * This is the slowpath switch to a preempted sandbox! * SIGUSR1 on the current thread and restore mcontext there! */ struct arch_context { reg_t regs[UREG_COUNT]; mcontext_t mctx; }; typedef struct arch_context arch_context_t; extern void __attribute__((noreturn)) worker_thread_sandbox_switch_preempt(void); extern __thread arch_context_t worker_thread_base_context; static void __attribute__((noinline)) arch_context_init(arch_context_t *actx, reg_t ip, reg_t sp) { memset(&actx->mctx, 0, sizeof(mcontext_t)); memset((void *)actx->regs, 0, sizeof(reg_t) * UREG_COUNT); if (sp) { /* * context_switch conventions: bp is expected to be on top of the stack * when co-op context switching.. * * so push sp on this new stack and use * that new sp as sp for switching to sandbox! */ asm volatile("movq %%rsp, %%rbx\n\t" "movq %%rax, %%rsp\n\t" "pushq %%rax\n\t" "movq %%rsp, %%rax\n\t" "movq %%rbx, %%rsp\n\t" : "=a"(sp) : "a"(sp) : "memory", "cc", "rbx"); } actx->regs[UREG_RSP] = sp; actx->regs[UREG_RIP] = ip; } /** * Preempt the current sandbox and start executing the next sandbox * @param mc - the context of the current thread of execution * @param ctx - the context that we want to restore * @return Return code in {0,1} * 0 = context restored successfully. * 1 = special processing because thread was last in a user-level context switch state */ static int arch_mcontext_restore(mcontext_t *mc, arch_context_t *ctx) { assert(ctx != &worker_thread_base_context); assert(!software_interrupt_is_enabled()); /* if ctx->regs[0] is set, this was last in a user-level context switch state! * else restore mcontext.. */ bool did_user_level_context_switch = ctx->regs[UREG_RSP]; if (did_user_level_context_switch) { mc->gregs[REG_RSP] = ctx->regs[UREG_RSP]; mc->gregs[REG_RIP] = ctx->regs[UREG_RIP] + ARCH_SIG_JMP_OFF; ctx->regs[UREG_RSP] = 0; return 1; } /* Restore mcontext */ memcpy(mc, &ctx->mctx, sizeof(mcontext_t)); memset(&ctx->mctx, 0, sizeof(mcontext_t)); return 0; } /** * Save the context of the currently executing process * @param ctx - destination * @param mc - source */ static void arch_mcontext_save(arch_context_t *ctx, mcontext_t *mc) { assert(ctx != &worker_thread_base_context); ctx->regs[UREG_RSP] = 0; memcpy(&ctx->mctx, mc, sizeof(mcontext_t)); } /** * @param current - the registers and context of the thing running * @param next - the registers and context of what we're switching to * @return always returns 0, indicating success * * NULL in either of these values indicates the "no sandbox to execute" state, * which defaults to resuming execution of main */ static inline int arch_context_switch(arch_context_t *current, arch_context_t *next) { /* if both current and next are NULL, there is no state change */ assert(current != NULL || next != NULL); /* Set any NULLs to worker_thread_base_context to resume execution of main */ if (current == NULL) current = &worker_thread_base_context; if (next == NULL) next = &worker_thread_base_context; reg_t *current_registers = current->regs, *next_registers = next->regs; assert(current_registers && next_registers); asm volatile("pushq %%rbp\n\t" "movq %%rsp, %%rbp\n\t" "movq $2f, 8(%%rax)\n\t" "movq %%rsp, (%%rax)\n\t" "cmpq $0, (%%rbx)\n\t" "je 1f\n\t" "movq (%%rbx), %%rsp\n\t" "jmpq *8(%%rbx)\n\t" "1:\n\t" "call worker_thread_sandbox_switch_preempt\n\t" ".align 8\n\t" "2:\n\t" "movq $0, (%%rbx)\n\t" ".align 8\n\t" "3:\n\t" "popq %%rbp\n\t" : : "a"(current_registers), "b"(next_registers) : "memory", "cc", "rcx", "rdx", "rsi", "rdi", "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15", "xmm0", "xmm1", "xmm2", "xmm3", "xmm4", "xmm5", "xmm6", "xmm7", "xmm8", "xmm9", "xmm10", "xmm11", "xmm12", "xmm13", "xmm14", "xmm15"); return 0; }