2801 lines
103 KiB
C
2801 lines
103 KiB
C
/*
|
|
* x86-64 signal handling routines
|
|
*
|
|
* Copyright 1999, 2005 Alexandre Julliard
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#if 0
|
|
#pragma makedep unix
|
|
#endif
|
|
|
|
#ifdef __x86_64__
|
|
|
|
#include "config.h"
|
|
#include "wine/port.h"
|
|
|
|
#include <assert.h>
|
|
#include <pthread.h>
|
|
#include <signal.h>
|
|
#include <stdlib.h>
|
|
#include <stdarg.h>
|
|
#include <stdio.h>
|
|
#include <sys/types.h>
|
|
#ifdef HAVE_UNISTD_H
|
|
# include <unistd.h>
|
|
#endif
|
|
#ifdef HAVE_MACHINE_SYSARCH_H
|
|
# include <machine/sysarch.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_PARAM_H
|
|
# include <sys/param.h>
|
|
#endif
|
|
#ifdef HAVE_SYSCALL_H
|
|
# include <syscall.h>
|
|
#else
|
|
# ifdef HAVE_SYS_SYSCALL_H
|
|
# include <sys/syscall.h>
|
|
# endif
|
|
#endif
|
|
#ifdef HAVE_SYS_SIGNAL_H
|
|
# include <sys/signal.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_UCONTEXT_H
|
|
# include <sys/ucontext.h>
|
|
#endif
|
|
#ifdef HAVE_LIBUNWIND
|
|
# define UNW_LOCAL_ONLY
|
|
# include <libunwind.h>
|
|
#endif
|
|
#ifdef __APPLE__
|
|
# include <mach/mach.h>
|
|
#endif
|
|
|
|
#define NONAMELESSUNION
|
|
#define NONAMELESSSTRUCT
|
|
#include "ntstatus.h"
|
|
#define WIN32_NO_STATUS
|
|
#include "windef.h"
|
|
#include "winternl.h"
|
|
#include "ddk/wdm.h"
|
|
#include "wine/exception.h"
|
|
#include "wine/list.h"
|
|
#include "wine/asm.h"
|
|
#include "unix_private.h"
|
|
#include "wine/debug.h"
|
|
|
|
WINE_DEFAULT_DEBUG_CHANNEL(seh);
|
|
|
|
/***********************************************************************
|
|
* signal context platform-specific definitions
|
|
*/
|
|
#ifdef linux
|
|
|
|
#include <asm/prctl.h>
|
|
static inline int arch_prctl( int func, void *ptr ) { return syscall( __NR_arch_prctl, func, ptr ); }
|
|
|
|
#ifndef FP_XSTATE_MAGIC1
|
|
#define FP_XSTATE_MAGIC1 0x46505853
|
|
#endif
|
|
|
|
#define RAX_sig(context) ((context)->uc_mcontext.gregs[REG_RAX])
|
|
#define RBX_sig(context) ((context)->uc_mcontext.gregs[REG_RBX])
|
|
#define RCX_sig(context) ((context)->uc_mcontext.gregs[REG_RCX])
|
|
#define RDX_sig(context) ((context)->uc_mcontext.gregs[REG_RDX])
|
|
#define RSI_sig(context) ((context)->uc_mcontext.gregs[REG_RSI])
|
|
#define RDI_sig(context) ((context)->uc_mcontext.gregs[REG_RDI])
|
|
#define RBP_sig(context) ((context)->uc_mcontext.gregs[REG_RBP])
|
|
#define R8_sig(context) ((context)->uc_mcontext.gregs[REG_R8])
|
|
#define R9_sig(context) ((context)->uc_mcontext.gregs[REG_R9])
|
|
#define R10_sig(context) ((context)->uc_mcontext.gregs[REG_R10])
|
|
#define R11_sig(context) ((context)->uc_mcontext.gregs[REG_R11])
|
|
#define R12_sig(context) ((context)->uc_mcontext.gregs[REG_R12])
|
|
#define R13_sig(context) ((context)->uc_mcontext.gregs[REG_R13])
|
|
#define R14_sig(context) ((context)->uc_mcontext.gregs[REG_R14])
|
|
#define R15_sig(context) ((context)->uc_mcontext.gregs[REG_R15])
|
|
#define CS_sig(context) (*((WORD *)&(context)->uc_mcontext.gregs[REG_CSGSFS] + 0))
|
|
#define GS_sig(context) (*((WORD *)&(context)->uc_mcontext.gregs[REG_CSGSFS] + 1))
|
|
#define FS_sig(context) (*((WORD *)&(context)->uc_mcontext.gregs[REG_CSGSFS] + 2))
|
|
#define RSP_sig(context) ((context)->uc_mcontext.gregs[REG_RSP])
|
|
#define RIP_sig(context) ((context)->uc_mcontext.gregs[REG_RIP])
|
|
#define EFL_sig(context) ((context)->uc_mcontext.gregs[REG_EFL])
|
|
#define TRAP_sig(context) ((context)->uc_mcontext.gregs[REG_TRAPNO])
|
|
#define ERROR_sig(context) ((context)->uc_mcontext.gregs[REG_ERR])
|
|
#define FPU_sig(context) ((XMM_SAVE_AREA32 *)((context)->uc_mcontext.fpregs))
|
|
#define XState_sig(fpu) (((unsigned int *)fpu->Reserved4)[12] == FP_XSTATE_MAGIC1 ? (XSTATE *)(fpu + 1) : NULL)
|
|
|
|
#elif defined(__FreeBSD__) || defined (__FreeBSD_kernel__)
|
|
|
|
#define RAX_sig(context) ((context)->uc_mcontext.mc_rax)
|
|
#define RBX_sig(context) ((context)->uc_mcontext.mc_rbx)
|
|
#define RCX_sig(context) ((context)->uc_mcontext.mc_rcx)
|
|
#define RDX_sig(context) ((context)->uc_mcontext.mc_rdx)
|
|
#define RSI_sig(context) ((context)->uc_mcontext.mc_rsi)
|
|
#define RDI_sig(context) ((context)->uc_mcontext.mc_rdi)
|
|
#define RBP_sig(context) ((context)->uc_mcontext.mc_rbp)
|
|
#define R8_sig(context) ((context)->uc_mcontext.mc_r8)
|
|
#define R9_sig(context) ((context)->uc_mcontext.mc_r9)
|
|
#define R10_sig(context) ((context)->uc_mcontext.mc_r10)
|
|
#define R11_sig(context) ((context)->uc_mcontext.mc_r11)
|
|
#define R12_sig(context) ((context)->uc_mcontext.mc_r12)
|
|
#define R13_sig(context) ((context)->uc_mcontext.mc_r13)
|
|
#define R14_sig(context) ((context)->uc_mcontext.mc_r14)
|
|
#define R15_sig(context) ((context)->uc_mcontext.mc_r15)
|
|
#define CS_sig(context) ((context)->uc_mcontext.mc_cs)
|
|
#define DS_sig(context) ((context)->uc_mcontext.mc_ds)
|
|
#define ES_sig(context) ((context)->uc_mcontext.mc_es)
|
|
#define FS_sig(context) ((context)->uc_mcontext.mc_fs)
|
|
#define GS_sig(context) ((context)->uc_mcontext.mc_gs)
|
|
#define SS_sig(context) ((context)->uc_mcontext.mc_ss)
|
|
#define EFL_sig(context) ((context)->uc_mcontext.mc_rflags)
|
|
#define RIP_sig(context) ((context)->uc_mcontext.mc_rip)
|
|
#define RSP_sig(context) ((context)->uc_mcontext.mc_rsp)
|
|
#define TRAP_sig(context) ((context)->uc_mcontext.mc_trapno)
|
|
#define ERROR_sig(context) ((context)->uc_mcontext.mc_err)
|
|
#define FPU_sig(context) ((XMM_SAVE_AREA32 *)((context)->uc_mcontext.mc_fpstate))
|
|
#define XState_sig(context) NULL
|
|
|
|
#elif defined(__NetBSD__)
|
|
|
|
#define RAX_sig(context) ((context)->uc_mcontext.__gregs[_REG_RAX])
|
|
#define RBX_sig(context) ((context)->uc_mcontext.__gregs[_REG_RBX])
|
|
#define RCX_sig(context) ((context)->uc_mcontext.__gregs[_REG_RCX])
|
|
#define RDX_sig(context) ((context)->uc_mcontext.__gregs[_REG_RDX])
|
|
#define RSI_sig(context) ((context)->uc_mcontext.__gregs[_REG_RSI])
|
|
#define RDI_sig(context) ((context)->uc_mcontext.__gregs[_REG_RDI])
|
|
#define RBP_sig(context) ((context)->uc_mcontext.__gregs[_REG_RBP])
|
|
#define R8_sig(context) ((context)->uc_mcontext.__gregs[_REG_R8])
|
|
#define R9_sig(context) ((context)->uc_mcontext.__gregs[_REG_R9])
|
|
#define R10_sig(context) ((context)->uc_mcontext.__gregs[_REG_R10])
|
|
#define R11_sig(context) ((context)->uc_mcontext.__gregs[_REG_R11])
|
|
#define R12_sig(context) ((context)->uc_mcontext.__gregs[_REG_R12])
|
|
#define R13_sig(context) ((context)->uc_mcontext.__gregs[_REG_R13])
|
|
#define R14_sig(context) ((context)->uc_mcontext.__gregs[_REG_R14])
|
|
#define R15_sig(context) ((context)->uc_mcontext.__gregs[_REG_R15])
|
|
#define CS_sig(context) ((context)->uc_mcontext.__gregs[_REG_CS])
|
|
#define DS_sig(context) ((context)->uc_mcontext.__gregs[_REG_DS])
|
|
#define ES_sig(context) ((context)->uc_mcontext.__gregs[_REG_ES])
|
|
#define FS_sig(context) ((context)->uc_mcontext.__gregs[_REG_FS])
|
|
#define GS_sig(context) ((context)->uc_mcontext.__gregs[_REG_GS])
|
|
#define SS_sig(context) ((context)->uc_mcontext.__gregs[_REG_SS])
|
|
#define EFL_sig(context) ((context)->uc_mcontext.__gregs[_REG_RFL])
|
|
#define RIP_sig(context) (*((unsigned long*)&(context)->uc_mcontext.__gregs[_REG_RIP]))
|
|
#define RSP_sig(context) (*((unsigned long*)&(context)->uc_mcontext.__gregs[_REG_URSP]))
|
|
#define TRAP_sig(context) ((context)->uc_mcontext.__gregs[_REG_TRAPNO])
|
|
#define ERROR_sig(context) ((context)->uc_mcontext.__gregs[_REG_ERR])
|
|
#define FPU_sig(context) ((XMM_SAVE_AREA32 *)((context)->uc_mcontext.__fpregs))
|
|
#define XState_sig(context) NULL
|
|
|
|
#elif defined (__APPLE__)
|
|
|
|
#define RAX_sig(context) ((context)->uc_mcontext->__ss.__rax)
|
|
#define RBX_sig(context) ((context)->uc_mcontext->__ss.__rbx)
|
|
#define RCX_sig(context) ((context)->uc_mcontext->__ss.__rcx)
|
|
#define RDX_sig(context) ((context)->uc_mcontext->__ss.__rdx)
|
|
#define RSI_sig(context) ((context)->uc_mcontext->__ss.__rsi)
|
|
#define RDI_sig(context) ((context)->uc_mcontext->__ss.__rdi)
|
|
#define RBP_sig(context) ((context)->uc_mcontext->__ss.__rbp)
|
|
#define R8_sig(context) ((context)->uc_mcontext->__ss.__r8)
|
|
#define R9_sig(context) ((context)->uc_mcontext->__ss.__r9)
|
|
#define R10_sig(context) ((context)->uc_mcontext->__ss.__r10)
|
|
#define R11_sig(context) ((context)->uc_mcontext->__ss.__r11)
|
|
#define R12_sig(context) ((context)->uc_mcontext->__ss.__r12)
|
|
#define R13_sig(context) ((context)->uc_mcontext->__ss.__r13)
|
|
#define R14_sig(context) ((context)->uc_mcontext->__ss.__r14)
|
|
#define R15_sig(context) ((context)->uc_mcontext->__ss.__r15)
|
|
#define CS_sig(context) ((context)->uc_mcontext->__ss.__cs)
|
|
#define FS_sig(context) ((context)->uc_mcontext->__ss.__fs)
|
|
#define GS_sig(context) ((context)->uc_mcontext->__ss.__gs)
|
|
#define EFL_sig(context) ((context)->uc_mcontext->__ss.__rflags)
|
|
#define RIP_sig(context) (*((unsigned long*)&(context)->uc_mcontext->__ss.__rip))
|
|
#define RSP_sig(context) (*((unsigned long*)&(context)->uc_mcontext->__ss.__rsp))
|
|
#define TRAP_sig(context) ((context)->uc_mcontext->__es.__trapno)
|
|
#define ERROR_sig(context) ((context)->uc_mcontext->__es.__err)
|
|
#define FPU_sig(context) ((XMM_SAVE_AREA32 *)&(context)->uc_mcontext->__fs.__fpu_fcw)
|
|
#define XState_sig(context) NULL
|
|
|
|
#else
|
|
#error You must define the signal context functions for your platform
|
|
#endif
|
|
|
|
enum i386_trap_code
|
|
{
|
|
TRAP_x86_DIVIDE = 0, /* Division by zero exception */
|
|
TRAP_x86_TRCTRAP = 1, /* Single-step exception */
|
|
TRAP_x86_NMI = 2, /* NMI interrupt */
|
|
TRAP_x86_BPTFLT = 3, /* Breakpoint exception */
|
|
TRAP_x86_OFLOW = 4, /* Overflow exception */
|
|
TRAP_x86_BOUND = 5, /* Bound range exception */
|
|
TRAP_x86_PRIVINFLT = 6, /* Invalid opcode exception */
|
|
TRAP_x86_DNA = 7, /* Device not available exception */
|
|
TRAP_x86_DOUBLEFLT = 8, /* Double fault exception */
|
|
TRAP_x86_FPOPFLT = 9, /* Coprocessor segment overrun */
|
|
TRAP_x86_TSSFLT = 10, /* Invalid TSS exception */
|
|
TRAP_x86_SEGNPFLT = 11, /* Segment not present exception */
|
|
TRAP_x86_STKFLT = 12, /* Stack fault */
|
|
TRAP_x86_PROTFLT = 13, /* General protection fault */
|
|
TRAP_x86_PAGEFLT = 14, /* Page fault */
|
|
TRAP_x86_ARITHTRAP = 16, /* Floating point exception */
|
|
TRAP_x86_ALIGNFLT = 17, /* Alignment check exception */
|
|
TRAP_x86_MCHK = 18, /* Machine check exception */
|
|
TRAP_x86_CACHEFLT = 19 /* Cache flush exception */
|
|
};
|
|
|
|
/* stack layout when calling an exception raise function */
|
|
struct stack_layout
|
|
{
|
|
CONTEXT context;
|
|
CONTEXT_EX context_ex;
|
|
EXCEPTION_RECORD rec;
|
|
ULONG64 rsi;
|
|
ULONG64 rdi;
|
|
ULONG64 rbp;
|
|
ULONG64 rip;
|
|
ULONG64 align;
|
|
char xstate[0]; /* If xstate is present it is allocated
|
|
* dynamically to provide 64 byte alignment. */
|
|
};
|
|
|
|
C_ASSERT((offsetof(struct stack_layout, xstate) == sizeof(struct stack_layout)));
|
|
|
|
C_ASSERT( sizeof(XSTATE) == 0x140 );
|
|
C_ASSERT( sizeof(struct stack_layout) == 0x5b0 ); /* Should match the size in call_user_exception_dispatcher(). */
|
|
|
|
struct syscall_frame
|
|
{
|
|
ULONG64 xmm[10 * 2]; /* xmm6-xmm15 */
|
|
ULONG64 mxcsr;
|
|
ULONG64 r12;
|
|
ULONG64 r13;
|
|
ULONG64 r14;
|
|
ULONG64 r15;
|
|
ULONG64 rdi;
|
|
ULONG64 rsi;
|
|
ULONG64 rbx;
|
|
ULONG64 rbp;
|
|
ULONG64 thunk_addr;
|
|
ULONG64 ret_addr;
|
|
};
|
|
|
|
struct amd64_thread_data
|
|
{
|
|
DWORD_PTR dr0; /* 02f0 debug registers */
|
|
DWORD_PTR dr1; /* 02f8 */
|
|
DWORD_PTR dr2; /* 0300 */
|
|
DWORD_PTR dr3; /* 0308 */
|
|
DWORD_PTR dr6; /* 0310 */
|
|
DWORD_PTR dr7; /* 0318 */
|
|
void *exit_frame; /* 0320 exit frame pointer */
|
|
struct syscall_frame *syscall_frame; /* 0328 syscall frame pointer */
|
|
};
|
|
|
|
C_ASSERT( sizeof(struct amd64_thread_data) <= sizeof(((struct ntdll_thread_data *)0)->cpu_data) );
|
|
C_ASSERT( offsetof( TEB, GdiTebBatch ) + offsetof( struct amd64_thread_data, exit_frame ) == 0x320 );
|
|
C_ASSERT( offsetof( TEB, GdiTebBatch ) + offsetof( struct amd64_thread_data, syscall_frame ) == 0x328 );
|
|
|
|
static inline struct amd64_thread_data *amd64_thread_data(void)
|
|
{
|
|
return (struct amd64_thread_data *)ntdll_get_thread_data()->cpu_data;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* Definitions for Dwarf unwind tables
|
|
*/
|
|
|
|
enum dwarf_call_frame_info
|
|
{
|
|
DW_CFA_advance_loc = 0x40,
|
|
DW_CFA_offset = 0x80,
|
|
DW_CFA_restore = 0xc0,
|
|
DW_CFA_nop = 0x00,
|
|
DW_CFA_set_loc = 0x01,
|
|
DW_CFA_advance_loc1 = 0x02,
|
|
DW_CFA_advance_loc2 = 0x03,
|
|
DW_CFA_advance_loc4 = 0x04,
|
|
DW_CFA_offset_extended = 0x05,
|
|
DW_CFA_restore_extended = 0x06,
|
|
DW_CFA_undefined = 0x07,
|
|
DW_CFA_same_value = 0x08,
|
|
DW_CFA_register = 0x09,
|
|
DW_CFA_remember_state = 0x0a,
|
|
DW_CFA_restore_state = 0x0b,
|
|
DW_CFA_def_cfa = 0x0c,
|
|
DW_CFA_def_cfa_register = 0x0d,
|
|
DW_CFA_def_cfa_offset = 0x0e,
|
|
DW_CFA_def_cfa_expression = 0x0f,
|
|
DW_CFA_expression = 0x10,
|
|
DW_CFA_offset_extended_sf = 0x11,
|
|
DW_CFA_def_cfa_sf = 0x12,
|
|
DW_CFA_def_cfa_offset_sf = 0x13,
|
|
DW_CFA_val_offset = 0x14,
|
|
DW_CFA_val_offset_sf = 0x15,
|
|
DW_CFA_val_expression = 0x16,
|
|
};
|
|
|
|
enum dwarf_operation
|
|
{
|
|
DW_OP_addr = 0x03,
|
|
DW_OP_deref = 0x06,
|
|
DW_OP_const1u = 0x08,
|
|
DW_OP_const1s = 0x09,
|
|
DW_OP_const2u = 0x0a,
|
|
DW_OP_const2s = 0x0b,
|
|
DW_OP_const4u = 0x0c,
|
|
DW_OP_const4s = 0x0d,
|
|
DW_OP_const8u = 0x0e,
|
|
DW_OP_const8s = 0x0f,
|
|
DW_OP_constu = 0x10,
|
|
DW_OP_consts = 0x11,
|
|
DW_OP_dup = 0x12,
|
|
DW_OP_drop = 0x13,
|
|
DW_OP_over = 0x14,
|
|
DW_OP_pick = 0x15,
|
|
DW_OP_swap = 0x16,
|
|
DW_OP_rot = 0x17,
|
|
DW_OP_xderef = 0x18,
|
|
DW_OP_abs = 0x19,
|
|
DW_OP_and = 0x1a,
|
|
DW_OP_div = 0x1b,
|
|
DW_OP_minus = 0x1c,
|
|
DW_OP_mod = 0x1d,
|
|
DW_OP_mul = 0x1e,
|
|
DW_OP_neg = 0x1f,
|
|
DW_OP_not = 0x20,
|
|
DW_OP_or = 0x21,
|
|
DW_OP_plus = 0x22,
|
|
DW_OP_plus_uconst = 0x23,
|
|
DW_OP_shl = 0x24,
|
|
DW_OP_shr = 0x25,
|
|
DW_OP_shra = 0x26,
|
|
DW_OP_xor = 0x27,
|
|
DW_OP_bra = 0x28,
|
|
DW_OP_eq = 0x29,
|
|
DW_OP_ge = 0x2a,
|
|
DW_OP_gt = 0x2b,
|
|
DW_OP_le = 0x2c,
|
|
DW_OP_lt = 0x2d,
|
|
DW_OP_ne = 0x2e,
|
|
DW_OP_skip = 0x2f,
|
|
DW_OP_lit0 = 0x30,
|
|
DW_OP_lit1 = 0x31,
|
|
DW_OP_lit2 = 0x32,
|
|
DW_OP_lit3 = 0x33,
|
|
DW_OP_lit4 = 0x34,
|
|
DW_OP_lit5 = 0x35,
|
|
DW_OP_lit6 = 0x36,
|
|
DW_OP_lit7 = 0x37,
|
|
DW_OP_lit8 = 0x38,
|
|
DW_OP_lit9 = 0x39,
|
|
DW_OP_lit10 = 0x3a,
|
|
DW_OP_lit11 = 0x3b,
|
|
DW_OP_lit12 = 0x3c,
|
|
DW_OP_lit13 = 0x3d,
|
|
DW_OP_lit14 = 0x3e,
|
|
DW_OP_lit15 = 0x3f,
|
|
DW_OP_lit16 = 0x40,
|
|
DW_OP_lit17 = 0x41,
|
|
DW_OP_lit18 = 0x42,
|
|
DW_OP_lit19 = 0x43,
|
|
DW_OP_lit20 = 0x44,
|
|
DW_OP_lit21 = 0x45,
|
|
DW_OP_lit22 = 0x46,
|
|
DW_OP_lit23 = 0x47,
|
|
DW_OP_lit24 = 0x48,
|
|
DW_OP_lit25 = 0x49,
|
|
DW_OP_lit26 = 0x4a,
|
|
DW_OP_lit27 = 0x4b,
|
|
DW_OP_lit28 = 0x4c,
|
|
DW_OP_lit29 = 0x4d,
|
|
DW_OP_lit30 = 0x4e,
|
|
DW_OP_lit31 = 0x4f,
|
|
DW_OP_reg0 = 0x50,
|
|
DW_OP_reg1 = 0x51,
|
|
DW_OP_reg2 = 0x52,
|
|
DW_OP_reg3 = 0x53,
|
|
DW_OP_reg4 = 0x54,
|
|
DW_OP_reg5 = 0x55,
|
|
DW_OP_reg6 = 0x56,
|
|
DW_OP_reg7 = 0x57,
|
|
DW_OP_reg8 = 0x58,
|
|
DW_OP_reg9 = 0x59,
|
|
DW_OP_reg10 = 0x5a,
|
|
DW_OP_reg11 = 0x5b,
|
|
DW_OP_reg12 = 0x5c,
|
|
DW_OP_reg13 = 0x5d,
|
|
DW_OP_reg14 = 0x5e,
|
|
DW_OP_reg15 = 0x5f,
|
|
DW_OP_reg16 = 0x60,
|
|
DW_OP_reg17 = 0x61,
|
|
DW_OP_reg18 = 0x62,
|
|
DW_OP_reg19 = 0x63,
|
|
DW_OP_reg20 = 0x64,
|
|
DW_OP_reg21 = 0x65,
|
|
DW_OP_reg22 = 0x66,
|
|
DW_OP_reg23 = 0x67,
|
|
DW_OP_reg24 = 0x68,
|
|
DW_OP_reg25 = 0x69,
|
|
DW_OP_reg26 = 0x6a,
|
|
DW_OP_reg27 = 0x6b,
|
|
DW_OP_reg28 = 0x6c,
|
|
DW_OP_reg29 = 0x6d,
|
|
DW_OP_reg30 = 0x6e,
|
|
DW_OP_reg31 = 0x6f,
|
|
DW_OP_breg0 = 0x70,
|
|
DW_OP_breg1 = 0x71,
|
|
DW_OP_breg2 = 0x72,
|
|
DW_OP_breg3 = 0x73,
|
|
DW_OP_breg4 = 0x74,
|
|
DW_OP_breg5 = 0x75,
|
|
DW_OP_breg6 = 0x76,
|
|
DW_OP_breg7 = 0x77,
|
|
DW_OP_breg8 = 0x78,
|
|
DW_OP_breg9 = 0x79,
|
|
DW_OP_breg10 = 0x7a,
|
|
DW_OP_breg11 = 0x7b,
|
|
DW_OP_breg12 = 0x7c,
|
|
DW_OP_breg13 = 0x7d,
|
|
DW_OP_breg14 = 0x7e,
|
|
DW_OP_breg15 = 0x7f,
|
|
DW_OP_breg16 = 0x80,
|
|
DW_OP_breg17 = 0x81,
|
|
DW_OP_breg18 = 0x82,
|
|
DW_OP_breg19 = 0x83,
|
|
DW_OP_breg20 = 0x84,
|
|
DW_OP_breg21 = 0x85,
|
|
DW_OP_breg22 = 0x86,
|
|
DW_OP_breg23 = 0x87,
|
|
DW_OP_breg24 = 0x88,
|
|
DW_OP_breg25 = 0x89,
|
|
DW_OP_breg26 = 0x8a,
|
|
DW_OP_breg27 = 0x8b,
|
|
DW_OP_breg28 = 0x8c,
|
|
DW_OP_breg29 = 0x8d,
|
|
DW_OP_breg30 = 0x8e,
|
|
DW_OP_breg31 = 0x8f,
|
|
DW_OP_regx = 0x90,
|
|
DW_OP_fbreg = 0x91,
|
|
DW_OP_bregx = 0x92,
|
|
DW_OP_piece = 0x93,
|
|
DW_OP_deref_size = 0x94,
|
|
DW_OP_xderef_size = 0x95,
|
|
DW_OP_nop = 0x96,
|
|
DW_OP_push_object_address = 0x97,
|
|
DW_OP_call2 = 0x98,
|
|
DW_OP_call4 = 0x99,
|
|
DW_OP_call_ref = 0x9a,
|
|
DW_OP_form_tls_address = 0x9b,
|
|
DW_OP_call_frame_cfa = 0x9c,
|
|
DW_OP_bit_piece = 0x9d,
|
|
DW_OP_lo_user = 0xe0,
|
|
DW_OP_hi_user = 0xff,
|
|
DW_OP_GNU_push_tls_address = 0xe0,
|
|
DW_OP_GNU_uninit = 0xf0,
|
|
DW_OP_GNU_encoded_addr = 0xf1,
|
|
};
|
|
|
|
#define DW_EH_PE_native 0x00
|
|
#define DW_EH_PE_leb128 0x01
|
|
#define DW_EH_PE_data2 0x02
|
|
#define DW_EH_PE_data4 0x03
|
|
#define DW_EH_PE_data8 0x04
|
|
#define DW_EH_PE_signed 0x08
|
|
#define DW_EH_PE_abs 0x00
|
|
#define DW_EH_PE_pcrel 0x10
|
|
#define DW_EH_PE_textrel 0x20
|
|
#define DW_EH_PE_datarel 0x30
|
|
#define DW_EH_PE_funcrel 0x40
|
|
#define DW_EH_PE_aligned 0x50
|
|
#define DW_EH_PE_indirect 0x80
|
|
#define DW_EH_PE_omit 0xff
|
|
|
|
struct dwarf_eh_bases
|
|
{
|
|
void *tbase;
|
|
void *dbase;
|
|
void *func;
|
|
};
|
|
|
|
struct dwarf_cie
|
|
{
|
|
unsigned int length;
|
|
int id;
|
|
unsigned char version;
|
|
unsigned char augmentation[1];
|
|
};
|
|
|
|
struct dwarf_fde
|
|
{
|
|
unsigned int length;
|
|
unsigned int cie_offset;
|
|
};
|
|
|
|
extern const struct dwarf_fde *_Unwind_Find_FDE (void *, struct dwarf_eh_bases *);
|
|
|
|
static unsigned char dwarf_get_u1( const unsigned char **p )
|
|
{
|
|
return *(*p)++;
|
|
}
|
|
|
|
static unsigned short dwarf_get_u2( const unsigned char **p )
|
|
{
|
|
unsigned int ret = (*p)[0] | ((*p)[1] << 8);
|
|
(*p) += 2;
|
|
return ret;
|
|
}
|
|
|
|
static unsigned int dwarf_get_u4( const unsigned char **p )
|
|
{
|
|
unsigned int ret = (*p)[0] | ((*p)[1] << 8) | ((*p)[2] << 16) | ((*p)[3] << 24);
|
|
(*p) += 4;
|
|
return ret;
|
|
}
|
|
|
|
static ULONG64 dwarf_get_u8( const unsigned char **p )
|
|
{
|
|
ULONG64 low = dwarf_get_u4( p );
|
|
ULONG64 high = dwarf_get_u4( p );
|
|
return low | (high << 32);
|
|
}
|
|
|
|
static ULONG_PTR dwarf_get_uleb128( const unsigned char **p )
|
|
{
|
|
ULONG_PTR ret = 0;
|
|
unsigned int shift = 0;
|
|
unsigned char byte;
|
|
|
|
do
|
|
{
|
|
byte = **p;
|
|
ret |= (ULONG_PTR)(byte & 0x7f) << shift;
|
|
shift += 7;
|
|
(*p)++;
|
|
} while (byte & 0x80);
|
|
return ret;
|
|
}
|
|
|
|
static LONG_PTR dwarf_get_sleb128( const unsigned char **p )
|
|
{
|
|
ULONG_PTR ret = 0;
|
|
unsigned int shift = 0;
|
|
unsigned char byte;
|
|
|
|
do
|
|
{
|
|
byte = **p;
|
|
ret |= (ULONG_PTR)(byte & 0x7f) << shift;
|
|
shift += 7;
|
|
(*p)++;
|
|
} while (byte & 0x80);
|
|
|
|
if ((shift < 8 * sizeof(ret)) && (byte & 0x40)) ret |= -((ULONG_PTR)1 << shift);
|
|
return ret;
|
|
}
|
|
|
|
static ULONG_PTR dwarf_get_ptr( const unsigned char **p, unsigned char encoding )
|
|
{
|
|
ULONG_PTR base;
|
|
|
|
if (encoding == DW_EH_PE_omit) return 0;
|
|
|
|
switch (encoding & 0xf0)
|
|
{
|
|
case DW_EH_PE_abs:
|
|
base = 0;
|
|
break;
|
|
case DW_EH_PE_pcrel:
|
|
base = (ULONG_PTR)*p;
|
|
break;
|
|
default:
|
|
FIXME( "unsupported encoding %02x\n", encoding );
|
|
return 0;
|
|
}
|
|
|
|
switch (encoding & 0x0f)
|
|
{
|
|
case DW_EH_PE_native:
|
|
return base + dwarf_get_u8( p );
|
|
case DW_EH_PE_leb128:
|
|
return base + dwarf_get_uleb128( p );
|
|
case DW_EH_PE_data2:
|
|
return base + dwarf_get_u2( p );
|
|
case DW_EH_PE_data4:
|
|
return base + dwarf_get_u4( p );
|
|
case DW_EH_PE_data8:
|
|
return base + dwarf_get_u8( p );
|
|
case DW_EH_PE_signed|DW_EH_PE_leb128:
|
|
return base + dwarf_get_sleb128( p );
|
|
case DW_EH_PE_signed|DW_EH_PE_data2:
|
|
return base + (signed short)dwarf_get_u2( p );
|
|
case DW_EH_PE_signed|DW_EH_PE_data4:
|
|
return base + (signed int)dwarf_get_u4( p );
|
|
case DW_EH_PE_signed|DW_EH_PE_data8:
|
|
return base + (LONG64)dwarf_get_u8( p );
|
|
default:
|
|
FIXME( "unsupported encoding %02x\n", encoding );
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
enum reg_rule
|
|
{
|
|
RULE_UNSET, /* not set at all */
|
|
RULE_UNDEFINED, /* undefined value */
|
|
RULE_SAME, /* same value as previous frame */
|
|
RULE_CFA_OFFSET, /* stored at cfa offset */
|
|
RULE_OTHER_REG, /* stored in other register */
|
|
RULE_EXPRESSION, /* address specified by expression */
|
|
RULE_VAL_EXPRESSION /* value specified by expression */
|
|
};
|
|
|
|
#define NB_FRAME_REGS 41
|
|
#define MAX_SAVED_STATES 16
|
|
|
|
struct frame_state
|
|
{
|
|
ULONG_PTR cfa_offset;
|
|
unsigned char cfa_reg;
|
|
enum reg_rule cfa_rule;
|
|
enum reg_rule rules[NB_FRAME_REGS];
|
|
ULONG64 regs[NB_FRAME_REGS];
|
|
};
|
|
|
|
struct frame_info
|
|
{
|
|
ULONG_PTR ip;
|
|
ULONG_PTR code_align;
|
|
LONG_PTR data_align;
|
|
unsigned char retaddr_reg;
|
|
unsigned char fde_encoding;
|
|
unsigned char signal_frame;
|
|
unsigned char state_sp;
|
|
struct frame_state state;
|
|
struct frame_state *state_stack;
|
|
};
|
|
|
|
static const char *dwarf_reg_names[NB_FRAME_REGS] =
|
|
{
|
|
/* 0-7 */ "%rax", "%rdx", "%rcx", "%rbx", "%rsi", "%rdi", "%rbp", "%rsp",
|
|
/* 8-16 */ "%r8", "%r9", "%r10", "%r11", "%r12", "%r13", "%r14", "%r15", "%rip",
|
|
/* 17-24 */ "%xmm0", "%xmm1", "%xmm2", "%xmm3", "%xmm4", "%xmm5", "%xmm6", "%xmm7",
|
|
/* 25-32 */ "%xmm8", "%xmm9", "%xmm10", "%xmm11", "%xmm12", "%xmm13", "%xmm14", "%xmm15",
|
|
/* 33-40 */ "%st0", "%st1", "%st2", "%st3", "%st4", "%st5", "%st6", "%st7"
|
|
};
|
|
|
|
static BOOL valid_reg( ULONG_PTR reg )
|
|
{
|
|
if (reg >= NB_FRAME_REGS) FIXME( "unsupported reg %lx\n", reg );
|
|
return (reg < NB_FRAME_REGS);
|
|
}
|
|
|
|
static void execute_cfa_instructions( const unsigned char *ptr, const unsigned char *end,
|
|
ULONG_PTR last_ip, struct frame_info *info )
|
|
{
|
|
while (ptr < end && info->ip < last_ip + info->signal_frame)
|
|
{
|
|
enum dwarf_call_frame_info op = *ptr++;
|
|
|
|
if (op & 0xc0)
|
|
{
|
|
switch (op & 0xc0)
|
|
{
|
|
case DW_CFA_advance_loc:
|
|
{
|
|
ULONG_PTR offset = (op & 0x3f) * info->code_align;
|
|
TRACE( "%lx: DW_CFA_advance_loc %lu\n", info->ip, offset );
|
|
info->ip += offset;
|
|
break;
|
|
}
|
|
case DW_CFA_offset:
|
|
{
|
|
ULONG_PTR reg = op & 0x3f;
|
|
LONG_PTR offset = dwarf_get_uleb128( &ptr ) * info->data_align;
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_offset %s, %ld\n", info->ip, dwarf_reg_names[reg], offset );
|
|
info->state.regs[reg] = offset;
|
|
info->state.rules[reg] = RULE_CFA_OFFSET;
|
|
break;
|
|
}
|
|
case DW_CFA_restore:
|
|
{
|
|
ULONG_PTR reg = op & 0x3f;
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_restore %s\n", info->ip, dwarf_reg_names[reg] );
|
|
info->state.rules[reg] = RULE_UNSET;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
else switch (op)
|
|
{
|
|
case DW_CFA_nop:
|
|
break;
|
|
case DW_CFA_set_loc:
|
|
{
|
|
ULONG_PTR loc = dwarf_get_ptr( &ptr, info->fde_encoding );
|
|
TRACE( "%lx: DW_CFA_set_loc %lx\n", info->ip, loc );
|
|
info->ip = loc;
|
|
break;
|
|
}
|
|
case DW_CFA_advance_loc1:
|
|
{
|
|
ULONG_PTR offset = *ptr++ * info->code_align;
|
|
TRACE( "%lx: DW_CFA_advance_loc1 %lu\n", info->ip, offset );
|
|
info->ip += offset;
|
|
break;
|
|
}
|
|
case DW_CFA_advance_loc2:
|
|
{
|
|
ULONG_PTR offset = dwarf_get_u2( &ptr ) * info->code_align;
|
|
TRACE( "%lx: DW_CFA_advance_loc2 %lu\n", info->ip, offset );
|
|
info->ip += offset;
|
|
break;
|
|
}
|
|
case DW_CFA_advance_loc4:
|
|
{
|
|
ULONG_PTR offset = dwarf_get_u4( &ptr ) * info->code_align;
|
|
TRACE( "%lx: DW_CFA_advance_loc4 %lu\n", info->ip, offset );
|
|
info->ip += offset;
|
|
break;
|
|
}
|
|
case DW_CFA_offset_extended:
|
|
case DW_CFA_offset_extended_sf:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
LONG_PTR offset = (op == DW_CFA_offset_extended) ? dwarf_get_uleb128( &ptr ) * info->data_align
|
|
: dwarf_get_sleb128( &ptr ) * info->data_align;
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_offset_extended %s, %ld\n", info->ip, dwarf_reg_names[reg], offset );
|
|
info->state.regs[reg] = offset;
|
|
info->state.rules[reg] = RULE_CFA_OFFSET;
|
|
break;
|
|
}
|
|
case DW_CFA_restore_extended:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_restore_extended %s\n", info->ip, dwarf_reg_names[reg] );
|
|
info->state.rules[reg] = RULE_UNSET;
|
|
break;
|
|
}
|
|
case DW_CFA_undefined:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_undefined %s\n", info->ip, dwarf_reg_names[reg] );
|
|
info->state.rules[reg] = RULE_UNDEFINED;
|
|
break;
|
|
}
|
|
case DW_CFA_same_value:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_same_value %s\n", info->ip, dwarf_reg_names[reg] );
|
|
info->state.regs[reg] = reg;
|
|
info->state.rules[reg] = RULE_SAME;
|
|
break;
|
|
}
|
|
case DW_CFA_register:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
ULONG_PTR reg2 = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg ) || !valid_reg( reg2 )) break;
|
|
TRACE( "%lx: DW_CFA_register %s == %s\n", info->ip, dwarf_reg_names[reg], dwarf_reg_names[reg2] );
|
|
info->state.regs[reg] = reg2;
|
|
info->state.rules[reg] = RULE_OTHER_REG;
|
|
break;
|
|
}
|
|
case DW_CFA_remember_state:
|
|
TRACE( "%lx: DW_CFA_remember_state\n", info->ip );
|
|
if (info->state_sp >= MAX_SAVED_STATES)
|
|
FIXME( "%lx: DW_CFA_remember_state too many nested saves\n", info->ip );
|
|
else
|
|
info->state_stack[info->state_sp++] = info->state;
|
|
break;
|
|
case DW_CFA_restore_state:
|
|
TRACE( "%lx: DW_CFA_restore_state\n", info->ip );
|
|
if (!info->state_sp)
|
|
FIXME( "%lx: DW_CFA_restore_state without corresponding save\n", info->ip );
|
|
else
|
|
info->state = info->state_stack[--info->state_sp];
|
|
break;
|
|
case DW_CFA_def_cfa:
|
|
case DW_CFA_def_cfa_sf:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
ULONG_PTR offset = (op == DW_CFA_def_cfa) ? dwarf_get_uleb128( &ptr )
|
|
: dwarf_get_sleb128( &ptr ) * info->data_align;
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_def_cfa %s, %lu\n", info->ip, dwarf_reg_names[reg], offset );
|
|
info->state.cfa_reg = reg;
|
|
info->state.cfa_offset = offset;
|
|
info->state.cfa_rule = RULE_CFA_OFFSET;
|
|
break;
|
|
}
|
|
case DW_CFA_def_cfa_register:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_def_cfa_register %s\n", info->ip, dwarf_reg_names[reg] );
|
|
info->state.cfa_reg = reg;
|
|
info->state.cfa_rule = RULE_CFA_OFFSET;
|
|
break;
|
|
}
|
|
case DW_CFA_def_cfa_offset:
|
|
case DW_CFA_def_cfa_offset_sf:
|
|
{
|
|
ULONG_PTR offset = (op == DW_CFA_def_cfa_offset) ? dwarf_get_uleb128( &ptr )
|
|
: dwarf_get_sleb128( &ptr ) * info->data_align;
|
|
TRACE( "%lx: DW_CFA_def_cfa_offset %lu\n", info->ip, offset );
|
|
info->state.cfa_offset = offset;
|
|
info->state.cfa_rule = RULE_CFA_OFFSET;
|
|
break;
|
|
}
|
|
case DW_CFA_def_cfa_expression:
|
|
{
|
|
ULONG_PTR expr = (ULONG_PTR)ptr;
|
|
ULONG_PTR len = dwarf_get_uleb128( &ptr );
|
|
TRACE( "%lx: DW_CFA_def_cfa_expression %lx-%lx\n", info->ip, expr, expr+len );
|
|
info->state.cfa_offset = expr;
|
|
info->state.cfa_rule = RULE_VAL_EXPRESSION;
|
|
ptr += len;
|
|
break;
|
|
}
|
|
case DW_CFA_expression:
|
|
case DW_CFA_val_expression:
|
|
{
|
|
ULONG_PTR reg = dwarf_get_uleb128( &ptr );
|
|
ULONG_PTR expr = (ULONG_PTR)ptr;
|
|
ULONG_PTR len = dwarf_get_uleb128( &ptr );
|
|
if (!valid_reg( reg )) break;
|
|
TRACE( "%lx: DW_CFA_%sexpression %s %lx-%lx\n",
|
|
info->ip, (op == DW_CFA_expression) ? "" : "val_", dwarf_reg_names[reg], expr, expr+len );
|
|
info->state.regs[reg] = expr;
|
|
info->state.rules[reg] = (op == DW_CFA_expression) ? RULE_EXPRESSION : RULE_VAL_EXPRESSION;
|
|
ptr += len;
|
|
break;
|
|
}
|
|
default:
|
|
FIXME( "%lx: unknown CFA opcode %02x\n", info->ip, op );
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* retrieve a context register from its dwarf number */
|
|
static void *get_context_reg( CONTEXT *context, ULONG_PTR dw_reg )
|
|
{
|
|
switch (dw_reg)
|
|
{
|
|
case 0: return &context->Rax;
|
|
case 1: return &context->Rdx;
|
|
case 2: return &context->Rcx;
|
|
case 3: return &context->Rbx;
|
|
case 4: return &context->Rsi;
|
|
case 5: return &context->Rdi;
|
|
case 6: return &context->Rbp;
|
|
case 7: return &context->Rsp;
|
|
case 8: return &context->R8;
|
|
case 9: return &context->R9;
|
|
case 10: return &context->R10;
|
|
case 11: return &context->R11;
|
|
case 12: return &context->R12;
|
|
case 13: return &context->R13;
|
|
case 14: return &context->R14;
|
|
case 15: return &context->R15;
|
|
case 16: return &context->Rip;
|
|
case 17: return &context->u.s.Xmm0;
|
|
case 18: return &context->u.s.Xmm1;
|
|
case 19: return &context->u.s.Xmm2;
|
|
case 20: return &context->u.s.Xmm3;
|
|
case 21: return &context->u.s.Xmm4;
|
|
case 22: return &context->u.s.Xmm5;
|
|
case 23: return &context->u.s.Xmm6;
|
|
case 24: return &context->u.s.Xmm7;
|
|
case 25: return &context->u.s.Xmm8;
|
|
case 26: return &context->u.s.Xmm9;
|
|
case 27: return &context->u.s.Xmm10;
|
|
case 28: return &context->u.s.Xmm11;
|
|
case 29: return &context->u.s.Xmm12;
|
|
case 30: return &context->u.s.Xmm13;
|
|
case 31: return &context->u.s.Xmm14;
|
|
case 32: return &context->u.s.Xmm15;
|
|
case 33: return &context->u.s.Legacy[0];
|
|
case 34: return &context->u.s.Legacy[1];
|
|
case 35: return &context->u.s.Legacy[2];
|
|
case 36: return &context->u.s.Legacy[3];
|
|
case 37: return &context->u.s.Legacy[4];
|
|
case 38: return &context->u.s.Legacy[5];
|
|
case 39: return &context->u.s.Legacy[6];
|
|
case 40: return &context->u.s.Legacy[7];
|
|
default: return NULL;
|
|
}
|
|
}
|
|
|
|
/* set a context register from its dwarf number */
|
|
static void set_context_reg( CONTEXT *context, ULONG_PTR dw_reg, void *val )
|
|
{
|
|
switch (dw_reg)
|
|
{
|
|
case 0: context->Rax = *(ULONG64 *)val; break;
|
|
case 1: context->Rdx = *(ULONG64 *)val; break;
|
|
case 2: context->Rcx = *(ULONG64 *)val; break;
|
|
case 3: context->Rbx = *(ULONG64 *)val; break;
|
|
case 4: context->Rsi = *(ULONG64 *)val; break;
|
|
case 5: context->Rdi = *(ULONG64 *)val; break;
|
|
case 6: context->Rbp = *(ULONG64 *)val; break;
|
|
case 7: context->Rsp = *(ULONG64 *)val; break;
|
|
case 8: context->R8 = *(ULONG64 *)val; break;
|
|
case 9: context->R9 = *(ULONG64 *)val; break;
|
|
case 10: context->R10 = *(ULONG64 *)val; break;
|
|
case 11: context->R11 = *(ULONG64 *)val; break;
|
|
case 12: context->R12 = *(ULONG64 *)val; break;
|
|
case 13: context->R13 = *(ULONG64 *)val; break;
|
|
case 14: context->R14 = *(ULONG64 *)val; break;
|
|
case 15: context->R15 = *(ULONG64 *)val; break;
|
|
case 16: context->Rip = *(ULONG64 *)val; break;
|
|
case 17: memcpy( &context->u.s.Xmm0, val, sizeof(M128A) ); break;
|
|
case 18: memcpy( &context->u.s.Xmm1, val, sizeof(M128A) ); break;
|
|
case 19: memcpy( &context->u.s.Xmm2, val, sizeof(M128A) ); break;
|
|
case 20: memcpy( &context->u.s.Xmm3, val, sizeof(M128A) ); break;
|
|
case 21: memcpy( &context->u.s.Xmm4, val, sizeof(M128A) ); break;
|
|
case 22: memcpy( &context->u.s.Xmm5, val, sizeof(M128A) ); break;
|
|
case 23: memcpy( &context->u.s.Xmm6, val, sizeof(M128A) ); break;
|
|
case 24: memcpy( &context->u.s.Xmm7, val, sizeof(M128A) ); break;
|
|
case 25: memcpy( &context->u.s.Xmm8, val, sizeof(M128A) ); break;
|
|
case 26: memcpy( &context->u.s.Xmm9, val, sizeof(M128A) ); break;
|
|
case 27: memcpy( &context->u.s.Xmm10, val, sizeof(M128A) ); break;
|
|
case 28: memcpy( &context->u.s.Xmm11, val, sizeof(M128A) ); break;
|
|
case 29: memcpy( &context->u.s.Xmm12, val, sizeof(M128A) ); break;
|
|
case 30: memcpy( &context->u.s.Xmm13, val, sizeof(M128A) ); break;
|
|
case 31: memcpy( &context->u.s.Xmm14, val, sizeof(M128A) ); break;
|
|
case 32: memcpy( &context->u.s.Xmm15, val, sizeof(M128A) ); break;
|
|
case 33: memcpy( &context->u.s.Legacy[0], val, sizeof(M128A) ); break;
|
|
case 34: memcpy( &context->u.s.Legacy[1], val, sizeof(M128A) ); break;
|
|
case 35: memcpy( &context->u.s.Legacy[2], val, sizeof(M128A) ); break;
|
|
case 36: memcpy( &context->u.s.Legacy[3], val, sizeof(M128A) ); break;
|
|
case 37: memcpy( &context->u.s.Legacy[4], val, sizeof(M128A) ); break;
|
|
case 38: memcpy( &context->u.s.Legacy[5], val, sizeof(M128A) ); break;
|
|
case 39: memcpy( &context->u.s.Legacy[6], val, sizeof(M128A) ); break;
|
|
case 40: memcpy( &context->u.s.Legacy[7], val, sizeof(M128A) ); break;
|
|
}
|
|
}
|
|
|
|
static ULONG_PTR eval_expression( const unsigned char *p, CONTEXT *context )
|
|
{
|
|
ULONG_PTR reg, tmp, stack[64];
|
|
int sp = -1;
|
|
ULONG_PTR len = dwarf_get_uleb128(&p);
|
|
const unsigned char *end = p + len;
|
|
|
|
while (p < end)
|
|
{
|
|
unsigned char opcode = dwarf_get_u1(&p);
|
|
|
|
if (opcode >= DW_OP_lit0 && opcode <= DW_OP_lit31)
|
|
stack[++sp] = opcode - DW_OP_lit0;
|
|
else if (opcode >= DW_OP_reg0 && opcode <= DW_OP_reg31)
|
|
stack[++sp] = *(ULONG_PTR *)get_context_reg( context, opcode - DW_OP_reg0 );
|
|
else if (opcode >= DW_OP_breg0 && opcode <= DW_OP_breg31)
|
|
stack[++sp] = *(ULONG_PTR *)get_context_reg( context, opcode - DW_OP_breg0 ) + dwarf_get_sleb128(&p);
|
|
else switch (opcode)
|
|
{
|
|
case DW_OP_nop: break;
|
|
case DW_OP_addr: stack[++sp] = dwarf_get_u8(&p); break;
|
|
case DW_OP_const1u: stack[++sp] = dwarf_get_u1(&p); break;
|
|
case DW_OP_const1s: stack[++sp] = (signed char)dwarf_get_u1(&p); break;
|
|
case DW_OP_const2u: stack[++sp] = dwarf_get_u2(&p); break;
|
|
case DW_OP_const2s: stack[++sp] = (short)dwarf_get_u2(&p); break;
|
|
case DW_OP_const4u: stack[++sp] = dwarf_get_u4(&p); break;
|
|
case DW_OP_const4s: stack[++sp] = (signed int)dwarf_get_u4(&p); break;
|
|
case DW_OP_const8u: stack[++sp] = dwarf_get_u8(&p); break;
|
|
case DW_OP_const8s: stack[++sp] = (LONG_PTR)dwarf_get_u8(&p); break;
|
|
case DW_OP_constu: stack[++sp] = dwarf_get_uleb128(&p); break;
|
|
case DW_OP_consts: stack[++sp] = dwarf_get_sleb128(&p); break;
|
|
case DW_OP_deref: stack[sp] = *(ULONG_PTR *)stack[sp]; break;
|
|
case DW_OP_dup: stack[sp + 1] = stack[sp]; sp++; break;
|
|
case DW_OP_drop: sp--; break;
|
|
case DW_OP_over: stack[sp + 1] = stack[sp - 1]; sp++; break;
|
|
case DW_OP_pick: stack[sp + 1] = stack[sp - dwarf_get_u1(&p)]; sp++; break;
|
|
case DW_OP_swap: tmp = stack[sp]; stack[sp] = stack[sp-1]; stack[sp-1] = tmp; break;
|
|
case DW_OP_rot: tmp = stack[sp]; stack[sp] = stack[sp-1]; stack[sp-1] = stack[sp-2]; stack[sp-2] = tmp; break;
|
|
case DW_OP_abs: stack[sp] = labs(stack[sp]); break;
|
|
case DW_OP_neg: stack[sp] = -stack[sp]; break;
|
|
case DW_OP_not: stack[sp] = ~stack[sp]; break;
|
|
case DW_OP_and: stack[sp-1] &= stack[sp]; sp--; break;
|
|
case DW_OP_or: stack[sp-1] |= stack[sp]; sp--; break;
|
|
case DW_OP_minus: stack[sp-1] -= stack[sp]; sp--; break;
|
|
case DW_OP_mul: stack[sp-1] *= stack[sp]; sp--; break;
|
|
case DW_OP_plus: stack[sp-1] += stack[sp]; sp--; break;
|
|
case DW_OP_xor: stack[sp-1] ^= stack[sp]; sp--; break;
|
|
case DW_OP_shl: stack[sp-1] <<= stack[sp]; sp--; break;
|
|
case DW_OP_shr: stack[sp-1] >>= stack[sp]; sp--; break;
|
|
case DW_OP_plus_uconst: stack[sp] += dwarf_get_uleb128(&p); break;
|
|
case DW_OP_shra: stack[sp-1] = (LONG_PTR)stack[sp-1] / (1 << stack[sp]); sp--; break;
|
|
case DW_OP_div: stack[sp-1] = (LONG_PTR)stack[sp-1] / (LONG_PTR)stack[sp]; sp--; break;
|
|
case DW_OP_mod: stack[sp-1] = (LONG_PTR)stack[sp-1] % (LONG_PTR)stack[sp]; sp--; break;
|
|
case DW_OP_ge: stack[sp-1] = ((LONG_PTR)stack[sp-1] >= (LONG_PTR)stack[sp]); sp--; break;
|
|
case DW_OP_gt: stack[sp-1] = ((LONG_PTR)stack[sp-1] > (LONG_PTR)stack[sp]); sp--; break;
|
|
case DW_OP_le: stack[sp-1] = ((LONG_PTR)stack[sp-1] <= (LONG_PTR)stack[sp]); sp--; break;
|
|
case DW_OP_lt: stack[sp-1] = ((LONG_PTR)stack[sp-1] < (LONG_PTR)stack[sp]); sp--; break;
|
|
case DW_OP_eq: stack[sp-1] = (stack[sp-1] == stack[sp]); sp--; break;
|
|
case DW_OP_ne: stack[sp-1] = (stack[sp-1] != stack[sp]); sp--; break;
|
|
case DW_OP_skip: tmp = (short)dwarf_get_u2(&p); p += tmp; break;
|
|
case DW_OP_bra: tmp = (short)dwarf_get_u2(&p); if (!stack[sp--]) p += tmp; break;
|
|
case DW_OP_GNU_encoded_addr: tmp = *p++; stack[++sp] = dwarf_get_ptr( &p, tmp ); break;
|
|
case DW_OP_regx: stack[++sp] = *(ULONG_PTR *)get_context_reg( context, dwarf_get_uleb128(&p) ); break;
|
|
case DW_OP_bregx:
|
|
reg = dwarf_get_uleb128(&p);
|
|
tmp = dwarf_get_sleb128(&p);
|
|
stack[++sp] = *(ULONG_PTR *)get_context_reg( context, reg ) + tmp;
|
|
break;
|
|
case DW_OP_deref_size:
|
|
switch (*p++)
|
|
{
|
|
case 1: stack[sp] = *(unsigned char *)stack[sp]; break;
|
|
case 2: stack[sp] = *(unsigned short *)stack[sp]; break;
|
|
case 4: stack[sp] = *(unsigned int *)stack[sp]; break;
|
|
case 8: stack[sp] = *(ULONG_PTR *)stack[sp]; break;
|
|
}
|
|
break;
|
|
default:
|
|
FIXME( "unhandled opcode %02x\n", opcode );
|
|
}
|
|
}
|
|
return stack[sp];
|
|
}
|
|
|
|
/* apply the computed frame info to the actual context */
|
|
static void apply_frame_state( CONTEXT *context, struct frame_state *state )
|
|
{
|
|
unsigned int i;
|
|
ULONG_PTR cfa, value;
|
|
CONTEXT new_context = *context;
|
|
|
|
switch (state->cfa_rule)
|
|
{
|
|
case RULE_EXPRESSION:
|
|
cfa = *(ULONG_PTR *)eval_expression( (const unsigned char *)state->cfa_offset, context );
|
|
break;
|
|
case RULE_VAL_EXPRESSION:
|
|
cfa = eval_expression( (const unsigned char *)state->cfa_offset, context );
|
|
break;
|
|
default:
|
|
cfa = *(ULONG_PTR *)get_context_reg( context, state->cfa_reg ) + state->cfa_offset;
|
|
break;
|
|
}
|
|
if (!cfa) return;
|
|
|
|
for (i = 0; i < NB_FRAME_REGS; i++)
|
|
{
|
|
switch (state->rules[i])
|
|
{
|
|
case RULE_UNSET:
|
|
case RULE_UNDEFINED:
|
|
case RULE_SAME:
|
|
break;
|
|
case RULE_CFA_OFFSET:
|
|
set_context_reg( &new_context, i, (char *)cfa + state->regs[i] );
|
|
break;
|
|
case RULE_OTHER_REG:
|
|
set_context_reg( &new_context, i, get_context_reg( context, state->regs[i] ));
|
|
break;
|
|
case RULE_EXPRESSION:
|
|
value = eval_expression( (const unsigned char *)state->regs[i], context );
|
|
set_context_reg( &new_context, i, (void *)value );
|
|
break;
|
|
case RULE_VAL_EXPRESSION:
|
|
value = eval_expression( (const unsigned char *)state->regs[i], context );
|
|
set_context_reg( &new_context, i, &value );
|
|
break;
|
|
}
|
|
}
|
|
new_context.Rsp = cfa;
|
|
*context = new_context;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* dwarf_virtual_unwind
|
|
*
|
|
* Equivalent of RtlVirtualUnwind for builtin modules.
|
|
*/
|
|
static NTSTATUS dwarf_virtual_unwind( ULONG64 ip, ULONG64 *frame,CONTEXT *context,
|
|
const struct dwarf_fde *fde, const struct dwarf_eh_bases *bases,
|
|
PEXCEPTION_ROUTINE *handler, void **handler_data )
|
|
{
|
|
const struct dwarf_cie *cie;
|
|
const unsigned char *ptr, *augmentation, *end;
|
|
ULONG_PTR len, code_end;
|
|
struct frame_info info;
|
|
struct frame_state state_stack[MAX_SAVED_STATES];
|
|
int aug_z_format = 0;
|
|
unsigned char lsda_encoding = DW_EH_PE_omit;
|
|
|
|
memset( &info, 0, sizeof(info) );
|
|
info.state_stack = state_stack;
|
|
info.ip = (ULONG_PTR)bases->func;
|
|
*handler = NULL;
|
|
|
|
cie = (const struct dwarf_cie *)((const char *)&fde->cie_offset - fde->cie_offset);
|
|
|
|
/* parse the CIE first */
|
|
|
|
if (cie->version != 1 && cie->version != 3)
|
|
{
|
|
FIXME( "unknown CIE version %u at %p\n", cie->version, cie );
|
|
return STATUS_INVALID_DISPOSITION;
|
|
}
|
|
ptr = cie->augmentation + strlen((const char *)cie->augmentation) + 1;
|
|
|
|
info.code_align = dwarf_get_uleb128( &ptr );
|
|
info.data_align = dwarf_get_sleb128( &ptr );
|
|
if (cie->version == 1)
|
|
info.retaddr_reg = *ptr++;
|
|
else
|
|
info.retaddr_reg = dwarf_get_uleb128( &ptr );
|
|
info.state.cfa_rule = RULE_CFA_OFFSET;
|
|
|
|
TRACE( "function %lx base %p cie %p len %x id %x version %x aug '%s' code_align %lu data_align %ld retaddr %s\n",
|
|
ip, bases->func, cie, cie->length, cie->id, cie->version, cie->augmentation,
|
|
info.code_align, info.data_align, dwarf_reg_names[info.retaddr_reg] );
|
|
|
|
end = NULL;
|
|
for (augmentation = cie->augmentation; *augmentation; augmentation++)
|
|
{
|
|
switch (*augmentation)
|
|
{
|
|
case 'z':
|
|
len = dwarf_get_uleb128( &ptr );
|
|
end = ptr + len;
|
|
aug_z_format = 1;
|
|
continue;
|
|
case 'L':
|
|
lsda_encoding = *ptr++;
|
|
continue;
|
|
case 'P':
|
|
{
|
|
unsigned char encoding = *ptr++;
|
|
*handler = (void *)dwarf_get_ptr( &ptr, encoding );
|
|
continue;
|
|
}
|
|
case 'R':
|
|
info.fde_encoding = *ptr++;
|
|
continue;
|
|
case 'S':
|
|
info.signal_frame = 1;
|
|
continue;
|
|
}
|
|
FIXME( "unknown augmentation '%c'\n", *augmentation );
|
|
if (!end) return STATUS_INVALID_DISPOSITION; /* cannot continue */
|
|
break;
|
|
}
|
|
if (end) ptr = end;
|
|
|
|
end = (const unsigned char *)(&cie->length + 1) + cie->length;
|
|
execute_cfa_instructions( ptr, end, ip, &info );
|
|
|
|
ptr = (const unsigned char *)(fde + 1);
|
|
info.ip = dwarf_get_ptr( &ptr, info.fde_encoding ); /* fde code start */
|
|
code_end = info.ip + dwarf_get_ptr( &ptr, info.fde_encoding & 0x0f ); /* fde code length */
|
|
|
|
if (aug_z_format) /* get length of augmentation data */
|
|
{
|
|
len = dwarf_get_uleb128( &ptr );
|
|
end = ptr + len;
|
|
}
|
|
else end = NULL;
|
|
|
|
*handler_data = (void *)dwarf_get_ptr( &ptr, lsda_encoding );
|
|
if (end) ptr = end;
|
|
|
|
end = (const unsigned char *)(&fde->length + 1) + fde->length;
|
|
TRACE( "fde %p len %x personality %p lsda %p code %lx-%lx\n",
|
|
fde, fde->length, *handler, *handler_data, info.ip, code_end );
|
|
execute_cfa_instructions( ptr, end, ip, &info );
|
|
*frame = context->Rsp;
|
|
apply_frame_state( context, &info.state );
|
|
|
|
TRACE( "next function rip=%016lx\n", context->Rip );
|
|
TRACE( " rax=%016lx rbx=%016lx rcx=%016lx rdx=%016lx\n",
|
|
context->Rax, context->Rbx, context->Rcx, context->Rdx );
|
|
TRACE( " rsi=%016lx rdi=%016lx rbp=%016lx rsp=%016lx\n",
|
|
context->Rsi, context->Rdi, context->Rbp, context->Rsp );
|
|
TRACE( " r8=%016lx r9=%016lx r10=%016lx r11=%016lx\n",
|
|
context->R8, context->R9, context->R10, context->R11 );
|
|
TRACE( " r12=%016lx r13=%016lx r14=%016lx r15=%016lx\n",
|
|
context->R12, context->R13, context->R14, context->R15 );
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
#ifdef HAVE_LIBUNWIND
|
|
/***********************************************************************
|
|
* libunwind_virtual_unwind
|
|
*
|
|
* Equivalent of RtlVirtualUnwind for builtin modules.
|
|
*/
|
|
static NTSTATUS libunwind_virtual_unwind( ULONG64 ip, ULONG64 *frame, CONTEXT *context,
|
|
PEXCEPTION_ROUTINE *handler, void **handler_data )
|
|
{
|
|
unw_context_t unw_context;
|
|
unw_cursor_t cursor;
|
|
unw_proc_info_t info;
|
|
int rc;
|
|
|
|
#ifdef __APPLE__
|
|
rc = unw_getcontext( &unw_context );
|
|
if (rc == UNW_ESUCCESS)
|
|
rc = unw_init_local( &cursor, &unw_context );
|
|
if (rc == UNW_ESUCCESS)
|
|
{
|
|
unw_set_reg( &cursor, UNW_REG_IP, context->Rip );
|
|
unw_set_reg( &cursor, UNW_REG_SP, context->Rsp );
|
|
unw_set_reg( &cursor, UNW_X86_64_RAX, context->Rax );
|
|
unw_set_reg( &cursor, UNW_X86_64_RDX, context->Rdx );
|
|
unw_set_reg( &cursor, UNW_X86_64_RCX, context->Rcx );
|
|
unw_set_reg( &cursor, UNW_X86_64_RBX, context->Rbx );
|
|
unw_set_reg( &cursor, UNW_X86_64_RSI, context->Rsi );
|
|
unw_set_reg( &cursor, UNW_X86_64_RDI, context->Rdi );
|
|
unw_set_reg( &cursor, UNW_X86_64_RBP, context->Rbp );
|
|
unw_set_reg( &cursor, UNW_X86_64_R8, context->R8 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R9, context->R9 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R10, context->R10 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R11, context->R11 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R12, context->R12 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R13, context->R13 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R14, context->R14 );
|
|
unw_set_reg( &cursor, UNW_X86_64_R15, context->R15 );
|
|
}
|
|
#else
|
|
RAX_sig(&unw_context) = context->Rax;
|
|
RCX_sig(&unw_context) = context->Rcx;
|
|
RDX_sig(&unw_context) = context->Rdx;
|
|
RBX_sig(&unw_context) = context->Rbx;
|
|
RSP_sig(&unw_context) = context->Rsp;
|
|
RBP_sig(&unw_context) = context->Rbp;
|
|
RSI_sig(&unw_context) = context->Rsi;
|
|
RDI_sig(&unw_context) = context->Rdi;
|
|
R8_sig(&unw_context) = context->R8;
|
|
R9_sig(&unw_context) = context->R9;
|
|
R10_sig(&unw_context) = context->R10;
|
|
R11_sig(&unw_context) = context->R11;
|
|
R12_sig(&unw_context) = context->R12;
|
|
R13_sig(&unw_context) = context->R13;
|
|
R14_sig(&unw_context) = context->R14;
|
|
R15_sig(&unw_context) = context->R15;
|
|
RIP_sig(&unw_context) = context->Rip;
|
|
CS_sig(&unw_context) = context->SegCs;
|
|
FS_sig(&unw_context) = context->SegFs;
|
|
GS_sig(&unw_context) = context->SegGs;
|
|
EFL_sig(&unw_context) = context->EFlags;
|
|
rc = unw_init_local( &cursor, &unw_context );
|
|
#endif
|
|
if (rc != UNW_ESUCCESS)
|
|
{
|
|
WARN( "setup failed: %d\n", rc );
|
|
return STATUS_INVALID_DISPOSITION;
|
|
}
|
|
|
|
*handler = NULL;
|
|
*frame = context->Rsp;
|
|
|
|
rc = unw_get_proc_info(&cursor, &info);
|
|
if (rc != UNW_ESUCCESS && rc != UNW_ENOINFO)
|
|
{
|
|
WARN( "failed to get info: %d\n", rc );
|
|
return STATUS_INVALID_DISPOSITION;
|
|
}
|
|
if (rc == UNW_ENOINFO || ip < info.start_ip || ip > info.end_ip || info.end_ip == info.start_ip + 1)
|
|
return STATUS_UNSUCCESSFUL;
|
|
|
|
TRACE( "ip %#lx function %#lx-%#lx personality %#lx lsda %#lx fde %#lx\n",
|
|
ip, (unsigned long)info.start_ip, (unsigned long)info.end_ip, (unsigned long)info.handler,
|
|
(unsigned long)info.lsda, (unsigned long)info.unwind_info );
|
|
|
|
if (!(rc = unw_step( &cursor )))
|
|
{
|
|
WARN( "last frame\n" );
|
|
return STATUS_UNSUCCESSFUL;
|
|
}
|
|
if (rc < 0)
|
|
{
|
|
WARN( "failed to unwind: %d\n", rc );
|
|
return STATUS_INVALID_DISPOSITION;
|
|
}
|
|
|
|
unw_get_reg( &cursor, UNW_REG_IP, (unw_word_t *)&context->Rip );
|
|
unw_get_reg( &cursor, UNW_REG_SP, (unw_word_t *)&context->Rsp );
|
|
unw_get_reg( &cursor, UNW_X86_64_RAX, (unw_word_t *)&context->Rax );
|
|
unw_get_reg( &cursor, UNW_X86_64_RDX, (unw_word_t *)&context->Rdx );
|
|
unw_get_reg( &cursor, UNW_X86_64_RCX, (unw_word_t *)&context->Rcx );
|
|
unw_get_reg( &cursor, UNW_X86_64_RBX, (unw_word_t *)&context->Rbx );
|
|
unw_get_reg( &cursor, UNW_X86_64_RSI, (unw_word_t *)&context->Rsi );
|
|
unw_get_reg( &cursor, UNW_X86_64_RDI, (unw_word_t *)&context->Rdi );
|
|
unw_get_reg( &cursor, UNW_X86_64_RBP, (unw_word_t *)&context->Rbp );
|
|
unw_get_reg( &cursor, UNW_X86_64_R8, (unw_word_t *)&context->R8 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R9, (unw_word_t *)&context->R9 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R10, (unw_word_t *)&context->R10 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R11, (unw_word_t *)&context->R11 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R12, (unw_word_t *)&context->R12 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R13, (unw_word_t *)&context->R13 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R14, (unw_word_t *)&context->R14 );
|
|
unw_get_reg( &cursor, UNW_X86_64_R15, (unw_word_t *)&context->R15 );
|
|
*handler = (void*)info.handler;
|
|
*handler_data = (void*)info.lsda;
|
|
|
|
TRACE( "next function rip=%016lx\n", context->Rip );
|
|
TRACE( " rax=%016lx rbx=%016lx rcx=%016lx rdx=%016lx\n",
|
|
context->Rax, context->Rbx, context->Rcx, context->Rdx );
|
|
TRACE( " rsi=%016lx rdi=%016lx rbp=%016lx rsp=%016lx\n",
|
|
context->Rsi, context->Rdi, context->Rbp, context->Rsp );
|
|
TRACE( " r8=%016lx r9=%016lx r10=%016lx r11=%016lx\n",
|
|
context->R8, context->R9, context->R10, context->R11 );
|
|
TRACE( " r12=%016lx r13=%016lx r14=%016lx r15=%016lx\n",
|
|
context->R12, context->R13, context->R14, context->R15 );
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
#endif
|
|
|
|
|
|
/***********************************************************************
|
|
* unwind_builtin_dll
|
|
*/
|
|
NTSTATUS CDECL unwind_builtin_dll( ULONG type, DISPATCHER_CONTEXT *dispatch, CONTEXT *context )
|
|
{
|
|
struct dwarf_eh_bases bases;
|
|
const struct dwarf_fde *fde = _Unwind_Find_FDE( (void *)(context->Rip - 1), &bases );
|
|
|
|
if (fde)
|
|
return dwarf_virtual_unwind( context->Rip, &dispatch->EstablisherFrame, context, fde,
|
|
&bases, &dispatch->LanguageHandler, &dispatch->HandlerData );
|
|
#ifdef HAVE_LIBUNWIND
|
|
return libunwind_virtual_unwind( context->Rip, &dispatch->EstablisherFrame, context,
|
|
&dispatch->LanguageHandler, &dispatch->HandlerData );
|
|
#endif
|
|
return STATUS_UNSUCCESSFUL;
|
|
}
|
|
|
|
|
|
static inline void set_sigcontext( const CONTEXT *context, ucontext_t *sigcontext )
|
|
{
|
|
RAX_sig(sigcontext) = context->Rax;
|
|
RCX_sig(sigcontext) = context->Rcx;
|
|
RDX_sig(sigcontext) = context->Rdx;
|
|
RBX_sig(sigcontext) = context->Rbx;
|
|
RSP_sig(sigcontext) = context->Rsp;
|
|
RBP_sig(sigcontext) = context->Rbp;
|
|
RSI_sig(sigcontext) = context->Rsi;
|
|
RDI_sig(sigcontext) = context->Rdi;
|
|
R8_sig(sigcontext) = context->R8;
|
|
R9_sig(sigcontext) = context->R9;
|
|
R10_sig(sigcontext) = context->R10;
|
|
R11_sig(sigcontext) = context->R11;
|
|
R12_sig(sigcontext) = context->R12;
|
|
R13_sig(sigcontext) = context->R13;
|
|
R14_sig(sigcontext) = context->R14;
|
|
R15_sig(sigcontext) = context->R15;
|
|
RIP_sig(sigcontext) = context->Rip;
|
|
CS_sig(sigcontext) = context->SegCs;
|
|
FS_sig(sigcontext) = context->SegFs;
|
|
GS_sig(sigcontext) = context->SegGs;
|
|
EFL_sig(sigcontext) = context->EFlags;
|
|
#ifdef DS_sig
|
|
DS_sig(sigcontext) = context->SegDs;
|
|
#endif
|
|
#ifdef ES_sig
|
|
ES_sig(sigcontext) = context->SegEs;
|
|
#endif
|
|
#ifdef SS_sig
|
|
SS_sig(sigcontext) = context->SegSs;
|
|
#endif
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* save_context
|
|
*
|
|
* Set the register values from a sigcontext.
|
|
*/
|
|
static void save_context( struct xcontext *xcontext, const ucontext_t *sigcontext )
|
|
{
|
|
CONTEXT *context = &xcontext->c;
|
|
|
|
context->ContextFlags = CONTEXT_CONTROL | CONTEXT_INTEGER | CONTEXT_SEGMENTS | CONTEXT_DEBUG_REGISTERS;
|
|
context->Rax = RAX_sig(sigcontext);
|
|
context->Rcx = RCX_sig(sigcontext);
|
|
context->Rdx = RDX_sig(sigcontext);
|
|
context->Rbx = RBX_sig(sigcontext);
|
|
context->Rsp = RSP_sig(sigcontext);
|
|
context->Rbp = RBP_sig(sigcontext);
|
|
context->Rsi = RSI_sig(sigcontext);
|
|
context->Rdi = RDI_sig(sigcontext);
|
|
context->R8 = R8_sig(sigcontext);
|
|
context->R9 = R9_sig(sigcontext);
|
|
context->R10 = R10_sig(sigcontext);
|
|
context->R11 = R11_sig(sigcontext);
|
|
context->R12 = R12_sig(sigcontext);
|
|
context->R13 = R13_sig(sigcontext);
|
|
context->R14 = R14_sig(sigcontext);
|
|
context->R15 = R15_sig(sigcontext);
|
|
context->Rip = RIP_sig(sigcontext);
|
|
context->SegCs = CS_sig(sigcontext);
|
|
context->SegFs = FS_sig(sigcontext);
|
|
context->SegGs = GS_sig(sigcontext);
|
|
context->EFlags = EFL_sig(sigcontext);
|
|
#ifdef DS_sig
|
|
context->SegDs = DS_sig(sigcontext);
|
|
#else
|
|
__asm__("movw %%ds,%0" : "=m" (context->SegDs));
|
|
#endif
|
|
#ifdef ES_sig
|
|
context->SegEs = ES_sig(sigcontext);
|
|
#else
|
|
__asm__("movw %%es,%0" : "=m" (context->SegEs));
|
|
#endif
|
|
#ifdef SS_sig
|
|
context->SegSs = SS_sig(sigcontext);
|
|
#else
|
|
__asm__("movw %%ss,%0" : "=m" (context->SegSs));
|
|
#endif
|
|
context->Dr0 = amd64_thread_data()->dr0;
|
|
context->Dr1 = amd64_thread_data()->dr1;
|
|
context->Dr2 = amd64_thread_data()->dr2;
|
|
context->Dr3 = amd64_thread_data()->dr3;
|
|
context->Dr6 = amd64_thread_data()->dr6;
|
|
context->Dr7 = amd64_thread_data()->dr7;
|
|
if (FPU_sig(sigcontext))
|
|
{
|
|
XSTATE *xs;
|
|
|
|
context->ContextFlags |= CONTEXT_FLOATING_POINT;
|
|
context->u.FltSave = *FPU_sig(sigcontext);
|
|
context->MxCsr = context->u.FltSave.MxCsr;
|
|
if ((xs = XState_sig(FPU_sig(sigcontext))))
|
|
{
|
|
/* xcontext and sigcontext are both on the signal stack, so we can
|
|
* just reference sigcontext without overflowing 32 bit XState.Offset */
|
|
context_init_xstate( context, xs );
|
|
assert( xcontext->c_ex.XState.Offset == (BYTE *)xs - (BYTE *)&xcontext->c_ex );
|
|
xcontext->host_compaction_mask = xs->CompactionMask;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* save_xstate
|
|
*
|
|
* Save the XState context
|
|
*/
|
|
static inline NTSTATUS save_xstate( CONTEXT *context )
|
|
{
|
|
CONTEXT_EX *context_ex = (CONTEXT_EX *)(context + 1);
|
|
DECLSPEC_ALIGN(64) struct
|
|
{
|
|
XSAVE_FORMAT xsave;
|
|
XSTATE xstate;
|
|
}
|
|
xsave_area;
|
|
XSTATE *xs;
|
|
|
|
if (!(user_shared_data->XState.EnabledFeatures && (xs = xstate_from_context( context ))))
|
|
return STATUS_SUCCESS;
|
|
|
|
if (context_ex->XState.Length < offsetof(XSTATE, YmmContext)
|
|
|| context_ex->XState.Length > sizeof(XSTATE))
|
|
return STATUS_INVALID_PARAMETER;
|
|
|
|
if (user_shared_data->XState.CompactionEnabled)
|
|
{
|
|
/* xsavec doesn't use anything from the save area. */
|
|
__asm__ volatile( "xsavec %0" : "=m"(xsave_area)
|
|
: "a" ((unsigned int)(xs->CompactionMask & (1 << XSTATE_AVX))), "d" (0) );
|
|
}
|
|
else
|
|
{
|
|
/* xsave preserves those bits in the mask which are not in EDX:EAX, so zero it. */
|
|
xsave_area.xstate.Mask = xsave_area.xstate.CompactionMask = 0;
|
|
__asm__ volatile( "xsave %0" : "=m"(xsave_area)
|
|
: "a" ((unsigned int)(xs->Mask & (1 << XSTATE_AVX))), "d" (0) );
|
|
}
|
|
|
|
memcpy(xs, &xsave_area.xstate, offsetof(XSTATE, YmmContext));
|
|
if (xs->Mask & (1 << XSTATE_AVX))
|
|
{
|
|
if (context_ex->XState.Length < sizeof(XSTATE))
|
|
return STATUS_BUFFER_OVERFLOW;
|
|
|
|
memcpy(&xs->YmmContext, &xsave_area.xstate.YmmContext, sizeof(xs->YmmContext));
|
|
}
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* restore_context
|
|
*
|
|
* Build a sigcontext from the register values.
|
|
*/
|
|
static void restore_context( const struct xcontext *xcontext, ucontext_t *sigcontext )
|
|
{
|
|
const CONTEXT *context = &xcontext->c;
|
|
XSTATE *xs;
|
|
|
|
amd64_thread_data()->dr0 = context->Dr0;
|
|
amd64_thread_data()->dr1 = context->Dr1;
|
|
amd64_thread_data()->dr2 = context->Dr2;
|
|
amd64_thread_data()->dr3 = context->Dr3;
|
|
amd64_thread_data()->dr6 = context->Dr6;
|
|
amd64_thread_data()->dr7 = context->Dr7;
|
|
set_sigcontext( context, sigcontext );
|
|
if (FPU_sig(sigcontext)) *FPU_sig(sigcontext) = context->u.FltSave;
|
|
if ((xs = XState_sig(FPU_sig(sigcontext))))
|
|
xs->CompactionMask = xcontext->host_compaction_mask;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* set_full_cpu_context
|
|
*
|
|
* Set the new CPU context.
|
|
*/
|
|
extern void set_full_cpu_context( const CONTEXT *context );
|
|
__ASM_GLOBAL_FUNC( set_full_cpu_context,
|
|
"subq $40,%rsp\n\t"
|
|
__ASM_SEH(".seh_stackalloc 0x40\n\t")
|
|
__ASM_SEH(".seh_endprologue\n\t")
|
|
__ASM_CFI(".cfi_adjust_cfa_offset 40\n\t")
|
|
"movq %gs:0x30,%rdx\n\t"
|
|
"movw 0x38(%rdi),%ax\n\t" /* context->SegCs */
|
|
"movq %rax,8(%rsp)\n\t"
|
|
"movw 0x42(%rdi),%ax\n\t" /* context->SegSs */
|
|
"movq %rax,32(%rsp)\n\t"
|
|
"movq 0x44(%rdi),%rax\n\t" /* context->Eflags */
|
|
"movq %rax,16(%rsp)\n\t"
|
|
"movq $0,0x328(%rdx)\n\t" /* amd64_thread_data()->syscall_frame */
|
|
"movq 0x80(%rdi),%rcx\n\t" /* context->Rcx */
|
|
"movq 0x88(%rdi),%rdx\n\t" /* context->Rdx */
|
|
"movq 0x90(%rdi),%rbx\n\t" /* context->Rbx */
|
|
"movq 0x98(%rdi),%rax\n\t" /* context->Rsp */
|
|
"movq %rax,24(%rsp)\n\t"
|
|
"movq 0xa0(%rdi),%rbp\n\t" /* context->Rbp */
|
|
"movq 0xa8(%rdi),%rsi\n\t" /* context->Rsi */
|
|
"movq 0xb8(%rdi),%r8\n\t" /* context->R8 */
|
|
"movq 0xc0(%rdi),%r9\n\t" /* context->R9 */
|
|
"movq 0xc8(%rdi),%r10\n\t" /* context->R10 */
|
|
"movq 0xd0(%rdi),%r11\n\t" /* context->R11 */
|
|
"movq 0xd8(%rdi),%r12\n\t" /* context->R12 */
|
|
"movq 0xe0(%rdi),%r13\n\t" /* context->R13 */
|
|
"movq 0xe8(%rdi),%r14\n\t" /* context->R14 */
|
|
"movq 0xf0(%rdi),%r15\n\t" /* context->R15 */
|
|
"movq 0xf8(%rdi),%rax\n\t" /* context->Rip */
|
|
"movq %rax,(%rsp)\n\t"
|
|
"fxrstor 0x100(%rdi)\n\t" /* context->FltSave */
|
|
"movq 0x78(%rdi),%rax\n\t" /* context->Rax */
|
|
"movq 0xb0(%rdi),%rdi\n\t" /* context->Rdi */
|
|
"iretq" );
|
|
|
|
|
|
/***********************************************************************
|
|
* restore_xstate
|
|
*
|
|
* Restore the XState context.
|
|
*/
|
|
static void restore_xstate( const CONTEXT *context )
|
|
{
|
|
XSAVE_FORMAT *xrstor_base;
|
|
XSTATE *xs;
|
|
|
|
if (!(xs = xstate_from_context( context )))
|
|
return;
|
|
|
|
xrstor_base = (XSAVE_FORMAT *)xs - 1;
|
|
|
|
if (!(xs->CompactionMask & ((ULONG64)1 << 63)))
|
|
{
|
|
/* Non-compacted xrstor will load Mxcsr regardless of the specified mask. Loading garbage there
|
|
* may lead to fault. We have only padding, no more used EXCEPTION_RECORD or unused context fields
|
|
* at the MxCsr restore location, so just put it there. */
|
|
assert( (void *)&xrstor_base->MxCsr > (void *)context->VectorRegister );
|
|
xrstor_base->MxCsr = context->u.FltSave.MxCsr;
|
|
xrstor_base->MxCsr_Mask = context->u.FltSave.MxCsr_Mask;
|
|
}
|
|
|
|
__asm__ volatile( "xrstor64 %0" : : "m"(*xrstor_base), "a" (4), "d" (0) );
|
|
}
|
|
|
|
/***********************************************************************
|
|
* get_server_context_flags
|
|
*
|
|
* Convert CPU-specific flags to generic server flags
|
|
*/
|
|
static unsigned int get_server_context_flags( DWORD flags )
|
|
{
|
|
unsigned int ret = 0;
|
|
|
|
flags &= ~CONTEXT_AMD64; /* get rid of CPU id */
|
|
if (flags & CONTEXT_CONTROL) ret |= SERVER_CTX_CONTROL;
|
|
if (flags & CONTEXT_INTEGER) ret |= SERVER_CTX_INTEGER;
|
|
if (flags & CONTEXT_SEGMENTS) ret |= SERVER_CTX_SEGMENTS;
|
|
if (flags & CONTEXT_FLOATING_POINT) ret |= SERVER_CTX_FLOATING_POINT;
|
|
if (flags & CONTEXT_DEBUG_REGISTERS) ret |= SERVER_CTX_DEBUG_REGISTERS;
|
|
if (flags & CONTEXT_XSTATE) ret |= SERVER_CTX_YMM_REGISTERS;
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* context_to_server
|
|
*
|
|
* Convert a register context to the server format.
|
|
*/
|
|
NTSTATUS context_to_server( context_t *to, const CONTEXT *from )
|
|
{
|
|
DWORD flags = from->ContextFlags & ~CONTEXT_AMD64; /* get rid of CPU id */
|
|
|
|
memset( to, 0, sizeof(*to) );
|
|
to->cpu = CPU_x86_64;
|
|
|
|
if (flags & CONTEXT_CONTROL)
|
|
{
|
|
to->flags |= SERVER_CTX_CONTROL;
|
|
to->ctl.x86_64_regs.rbp = from->Rbp;
|
|
to->ctl.x86_64_regs.rip = from->Rip;
|
|
to->ctl.x86_64_regs.rsp = from->Rsp;
|
|
to->ctl.x86_64_regs.cs = from->SegCs;
|
|
to->ctl.x86_64_regs.ss = from->SegSs;
|
|
to->ctl.x86_64_regs.flags = from->EFlags;
|
|
}
|
|
if (flags & CONTEXT_INTEGER)
|
|
{
|
|
to->flags |= SERVER_CTX_INTEGER;
|
|
to->integer.x86_64_regs.rax = from->Rax;
|
|
to->integer.x86_64_regs.rcx = from->Rcx;
|
|
to->integer.x86_64_regs.rdx = from->Rdx;
|
|
to->integer.x86_64_regs.rbx = from->Rbx;
|
|
to->integer.x86_64_regs.rsi = from->Rsi;
|
|
to->integer.x86_64_regs.rdi = from->Rdi;
|
|
to->integer.x86_64_regs.r8 = from->R8;
|
|
to->integer.x86_64_regs.r9 = from->R9;
|
|
to->integer.x86_64_regs.r10 = from->R10;
|
|
to->integer.x86_64_regs.r11 = from->R11;
|
|
to->integer.x86_64_regs.r12 = from->R12;
|
|
to->integer.x86_64_regs.r13 = from->R13;
|
|
to->integer.x86_64_regs.r14 = from->R14;
|
|
to->integer.x86_64_regs.r15 = from->R15;
|
|
}
|
|
if (flags & CONTEXT_SEGMENTS)
|
|
{
|
|
to->flags |= SERVER_CTX_SEGMENTS;
|
|
to->seg.x86_64_regs.ds = from->SegDs;
|
|
to->seg.x86_64_regs.es = from->SegEs;
|
|
to->seg.x86_64_regs.fs = from->SegFs;
|
|
to->seg.x86_64_regs.gs = from->SegGs;
|
|
}
|
|
if (flags & CONTEXT_FLOATING_POINT)
|
|
{
|
|
to->flags |= SERVER_CTX_FLOATING_POINT;
|
|
memcpy( to->fp.x86_64_regs.fpregs, &from->u.FltSave, sizeof(to->fp.x86_64_regs.fpregs) );
|
|
}
|
|
if (flags & CONTEXT_DEBUG_REGISTERS)
|
|
{
|
|
to->flags |= SERVER_CTX_DEBUG_REGISTERS;
|
|
to->debug.x86_64_regs.dr0 = from->Dr0;
|
|
to->debug.x86_64_regs.dr1 = from->Dr1;
|
|
to->debug.x86_64_regs.dr2 = from->Dr2;
|
|
to->debug.x86_64_regs.dr3 = from->Dr3;
|
|
to->debug.x86_64_regs.dr6 = from->Dr6;
|
|
to->debug.x86_64_regs.dr7 = from->Dr7;
|
|
}
|
|
xstate_to_server( to, xstate_from_context( from ) );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* context_from_server
|
|
*
|
|
* Convert a register context from the server format.
|
|
*/
|
|
NTSTATUS context_from_server( CONTEXT *to, const context_t *from )
|
|
{
|
|
if (from->cpu != CPU_x86_64) return STATUS_INVALID_PARAMETER;
|
|
|
|
to->ContextFlags = CONTEXT_AMD64 | (to->ContextFlags & 0x40);
|
|
if (from->flags & SERVER_CTX_CONTROL)
|
|
{
|
|
to->ContextFlags |= CONTEXT_CONTROL;
|
|
to->Rbp = from->ctl.x86_64_regs.rbp;
|
|
to->Rip = from->ctl.x86_64_regs.rip;
|
|
to->Rsp = from->ctl.x86_64_regs.rsp;
|
|
to->SegCs = from->ctl.x86_64_regs.cs;
|
|
to->SegSs = from->ctl.x86_64_regs.ss;
|
|
to->EFlags = from->ctl.x86_64_regs.flags;
|
|
}
|
|
|
|
if (from->flags & SERVER_CTX_INTEGER)
|
|
{
|
|
to->ContextFlags |= CONTEXT_INTEGER;
|
|
to->Rax = from->integer.x86_64_regs.rax;
|
|
to->Rcx = from->integer.x86_64_regs.rcx;
|
|
to->Rdx = from->integer.x86_64_regs.rdx;
|
|
to->Rbx = from->integer.x86_64_regs.rbx;
|
|
to->Rsi = from->integer.x86_64_regs.rsi;
|
|
to->Rdi = from->integer.x86_64_regs.rdi;
|
|
to->R8 = from->integer.x86_64_regs.r8;
|
|
to->R9 = from->integer.x86_64_regs.r9;
|
|
to->R10 = from->integer.x86_64_regs.r10;
|
|
to->R11 = from->integer.x86_64_regs.r11;
|
|
to->R12 = from->integer.x86_64_regs.r12;
|
|
to->R13 = from->integer.x86_64_regs.r13;
|
|
to->R14 = from->integer.x86_64_regs.r14;
|
|
to->R15 = from->integer.x86_64_regs.r15;
|
|
}
|
|
if (from->flags & SERVER_CTX_SEGMENTS)
|
|
{
|
|
to->ContextFlags |= CONTEXT_SEGMENTS;
|
|
to->SegDs = from->seg.x86_64_regs.ds;
|
|
to->SegEs = from->seg.x86_64_regs.es;
|
|
to->SegFs = from->seg.x86_64_regs.fs;
|
|
to->SegGs = from->seg.x86_64_regs.gs;
|
|
}
|
|
if (from->flags & SERVER_CTX_FLOATING_POINT)
|
|
{
|
|
to->ContextFlags |= CONTEXT_FLOATING_POINT;
|
|
memcpy( &to->u.FltSave, from->fp.x86_64_regs.fpregs, sizeof(from->fp.x86_64_regs.fpregs) );
|
|
to->MxCsr = to->u.FltSave.MxCsr;
|
|
}
|
|
if (from->flags & SERVER_CTX_DEBUG_REGISTERS)
|
|
{
|
|
to->ContextFlags |= CONTEXT_DEBUG_REGISTERS;
|
|
to->Dr0 = from->debug.x86_64_regs.dr0;
|
|
to->Dr1 = from->debug.x86_64_regs.dr1;
|
|
to->Dr2 = from->debug.x86_64_regs.dr2;
|
|
to->Dr3 = from->debug.x86_64_regs.dr3;
|
|
to->Dr6 = from->debug.x86_64_regs.dr6;
|
|
to->Dr7 = from->debug.x86_64_regs.dr7;
|
|
}
|
|
xstate_from_server( xstate_from_context( to ), from );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtSetContextThread (NTDLL.@)
|
|
* ZwSetContextThread (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtSetContextThread( HANDLE handle, const CONTEXT *context )
|
|
{
|
|
NTSTATUS ret = STATUS_SUCCESS;
|
|
DWORD flags = context->ContextFlags & ~CONTEXT_AMD64;
|
|
BOOL self = (handle == GetCurrentThread());
|
|
|
|
/* debug registers require a server call */
|
|
if (self && (flags & CONTEXT_DEBUG_REGISTERS))
|
|
self = (amd64_thread_data()->dr0 == context->Dr0 &&
|
|
amd64_thread_data()->dr1 == context->Dr1 &&
|
|
amd64_thread_data()->dr2 == context->Dr2 &&
|
|
amd64_thread_data()->dr3 == context->Dr3 &&
|
|
amd64_thread_data()->dr6 == context->Dr6 &&
|
|
amd64_thread_data()->dr7 == context->Dr7);
|
|
|
|
if (!self)
|
|
{
|
|
context_t server_context;
|
|
|
|
context_to_server( &server_context, context );
|
|
ret = set_thread_context( handle, &server_context, &self );
|
|
if (ret || !self) return ret;
|
|
if (flags & CONTEXT_DEBUG_REGISTERS)
|
|
{
|
|
amd64_thread_data()->dr0 = context->Dr0;
|
|
amd64_thread_data()->dr1 = context->Dr1;
|
|
amd64_thread_data()->dr2 = context->Dr2;
|
|
amd64_thread_data()->dr3 = context->Dr3;
|
|
amd64_thread_data()->dr6 = context->Dr6;
|
|
amd64_thread_data()->dr7 = context->Dr7;
|
|
}
|
|
}
|
|
|
|
restore_xstate( context );
|
|
|
|
if (flags & CONTEXT_FULL)
|
|
{
|
|
if (!(flags & CONTEXT_CONTROL))
|
|
FIXME( "setting partial context (%x) not supported\n", flags );
|
|
else
|
|
set_full_cpu_context( context );
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtGetContextThread (NTDLL.@)
|
|
* ZwGetContextThread (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtGetContextThread( HANDLE handle, CONTEXT *context )
|
|
{
|
|
NTSTATUS ret, xsave_status;
|
|
DWORD needed_flags;
|
|
struct syscall_frame *frame = amd64_thread_data()->syscall_frame;
|
|
BOOL self = (handle == GetCurrentThread());
|
|
|
|
if (!context) return STATUS_INVALID_PARAMETER;
|
|
|
|
/* Save xstate before any calls which can potentially change volatile ymm registers.
|
|
* E. g., debug output will clobber ymm registers. */
|
|
xsave_status = self ? save_xstate( context ) : STATUS_SUCCESS;
|
|
|
|
needed_flags = context->ContextFlags & ~CONTEXT_AMD64;
|
|
|
|
/* debug registers require a server call */
|
|
if (context->ContextFlags & (CONTEXT_DEBUG_REGISTERS & ~CONTEXT_AMD64)) self = FALSE;
|
|
|
|
if (!self)
|
|
{
|
|
context_t server_context;
|
|
unsigned int server_flags = get_server_context_flags( context->ContextFlags );
|
|
|
|
if ((ret = get_thread_context( handle, &server_context, server_flags, &self ))) return ret;
|
|
if ((ret = context_from_server( context, &server_context ))) return ret;
|
|
needed_flags &= ~context->ContextFlags;
|
|
}
|
|
|
|
if (self)
|
|
{
|
|
if (needed_flags & CONTEXT_INTEGER)
|
|
{
|
|
context->Rax = 0;
|
|
context->Rbx = frame->rbx;
|
|
context->Rcx = 0;
|
|
context->Rdx = 0;
|
|
context->Rsi = frame->rsi;
|
|
context->Rdi = frame->rdi;
|
|
context->R8 = 0;
|
|
context->R9 = 0;
|
|
context->R10 = 0;
|
|
context->R11 = 0;
|
|
context->R12 = frame->r12;
|
|
context->R13 = frame->r13;
|
|
context->R14 = frame->r14;
|
|
context->R15 = frame->r15;
|
|
context->ContextFlags |= CONTEXT_INTEGER;
|
|
}
|
|
if (needed_flags & CONTEXT_CONTROL)
|
|
{
|
|
context->Rsp = (ULONG64)&frame->ret_addr;
|
|
context->Rbp = frame->rbp;
|
|
context->Rip = frame->thunk_addr;
|
|
context->EFlags = 0x202;
|
|
__asm__( "movw %%cs,%0" : "=g" (context->SegCs) );
|
|
__asm__( "movw %%ss,%0" : "=g" (context->SegSs) );
|
|
context->ContextFlags |= CONTEXT_CONTROL;
|
|
}
|
|
if (needed_flags & CONTEXT_SEGMENTS)
|
|
{
|
|
__asm__( "movw %%ds,%0" : "=g" (context->SegDs) );
|
|
__asm__( "movw %%es,%0" : "=g" (context->SegEs) );
|
|
__asm__( "movw %%fs,%0" : "=g" (context->SegFs) );
|
|
__asm__( "movw %%gs,%0" : "=g" (context->SegGs) );
|
|
context->ContextFlags |= CONTEXT_SEGMENTS;
|
|
}
|
|
if (needed_flags & CONTEXT_FLOATING_POINT)
|
|
{
|
|
__asm__( "fxsave %0" : "=m" (context->u.FltSave) );
|
|
context->MxCsr = frame->mxcsr;
|
|
memset( &context->u.s.Xmm0, 0, 6 * sizeof(context->u.s.Xmm0) );
|
|
memcpy( &context->u.s.Xmm6, frame->xmm, 10 * sizeof(context->u.s.Xmm0) );
|
|
context->ContextFlags |= CONTEXT_FLOATING_POINT;
|
|
}
|
|
/* update the cached version of the debug registers */
|
|
if (context->ContextFlags & (CONTEXT_DEBUG_REGISTERS & ~CONTEXT_AMD64))
|
|
{
|
|
amd64_thread_data()->dr0 = context->Dr0;
|
|
amd64_thread_data()->dr1 = context->Dr1;
|
|
amd64_thread_data()->dr2 = context->Dr2;
|
|
amd64_thread_data()->dr3 = context->Dr3;
|
|
amd64_thread_data()->dr6 = context->Dr6;
|
|
amd64_thread_data()->dr7 = context->Dr7;
|
|
}
|
|
}
|
|
|
|
return xsave_status;
|
|
}
|
|
|
|
extern void CDECL raise_func_trampoline( void *dispatcher );
|
|
|
|
__ASM_GLOBAL_FUNC( raise_func_trampoline,
|
|
"jmpq *%r8\n\t")
|
|
|
|
/***********************************************************************
|
|
* setup_raise_exception
|
|
*/
|
|
static void setup_raise_exception( ucontext_t *sigcontext, EXCEPTION_RECORD *rec, struct xcontext *xcontext )
|
|
{
|
|
void *stack_ptr = (void *)(RSP_sig(sigcontext) & ~15);
|
|
CONTEXT *context = &xcontext->c;
|
|
struct stack_layout *stack;
|
|
size_t stack_size;
|
|
NTSTATUS status;
|
|
XSTATE *src_xs;
|
|
|
|
if (rec->ExceptionCode == EXCEPTION_SINGLE_STEP)
|
|
{
|
|
/* when single stepping can't tell whether this is a hw bp or a
|
|
* single step interrupt. try to avoid as much overhead as possible
|
|
* and only do a server call if there is any hw bp enabled. */
|
|
|
|
if (!(context->EFlags & 0x100) || (context->Dr7 & 0xff))
|
|
{
|
|
/* (possible) hardware breakpoint, fetch the debug registers */
|
|
DWORD saved_flags = context->ContextFlags;
|
|
context->ContextFlags = CONTEXT_DEBUG_REGISTERS;
|
|
NtGetContextThread(GetCurrentThread(), context);
|
|
context->ContextFlags |= saved_flags; /* restore flags */
|
|
}
|
|
context->EFlags &= ~0x100; /* clear single-step flag */
|
|
}
|
|
|
|
status = send_debug_event( rec, context, TRUE );
|
|
if (status == DBG_CONTINUE || status == DBG_EXCEPTION_HANDLED)
|
|
{
|
|
restore_context( xcontext, sigcontext );
|
|
return;
|
|
}
|
|
|
|
/* fix up instruction pointer in context for EXCEPTION_BREAKPOINT */
|
|
if (rec->ExceptionCode == EXCEPTION_BREAKPOINT) context->Rip--;
|
|
|
|
stack_size = sizeof(*stack);
|
|
if ((src_xs = xstate_from_context( context )))
|
|
{
|
|
stack_size += (ULONG_PTR)stack_ptr - (((ULONG_PTR)stack_ptr
|
|
- sizeof(XSTATE)) & ~(ULONG_PTR)63);
|
|
}
|
|
|
|
stack = virtual_setup_exception( stack_ptr, stack_size, rec );
|
|
stack->rec = *rec;
|
|
stack->context = *context;
|
|
if (src_xs)
|
|
{
|
|
XSTATE *dst_xs = (XSTATE *)stack->xstate;
|
|
|
|
assert( !((ULONG_PTR)dst_xs & 63) );
|
|
context_init_xstate( &stack->context, stack->xstate );
|
|
memset( dst_xs, 0, offsetof(XSTATE, YmmContext) );
|
|
dst_xs->CompactionMask = user_shared_data->XState.CompactionEnabled ? 0x8000000000000004 : 0;
|
|
if (src_xs->Mask & 4)
|
|
{
|
|
dst_xs->Mask = 4;
|
|
memcpy( &dst_xs->YmmContext, &src_xs->YmmContext, sizeof(dst_xs->YmmContext) );
|
|
}
|
|
}
|
|
|
|
RIP_sig(sigcontext) = (ULONG_PTR)raise_func_trampoline;
|
|
R8_sig(sigcontext) = (ULONG_PTR)pKiUserExceptionDispatcher;
|
|
RSP_sig(sigcontext) = (ULONG_PTR)stack;
|
|
/* clear single-step, direction, and align check flag */
|
|
EFL_sig(sigcontext) &= ~(0x100|0x400|0x40000);
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* setup_exception
|
|
*
|
|
* Setup a proper stack frame for the raise function, and modify the
|
|
* sigcontext so that the return from the signal handler will call
|
|
* the raise function.
|
|
*/
|
|
static void setup_exception( ucontext_t *sigcontext, EXCEPTION_RECORD *rec )
|
|
{
|
|
struct xcontext context;
|
|
|
|
rec->ExceptionAddress = (void *)RIP_sig(sigcontext);
|
|
save_context( &context, sigcontext );
|
|
setup_raise_exception( sigcontext, rec, &context );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* call_user_apc_dispatcher
|
|
*/
|
|
__ASM_GLOBAL_FUNC( call_user_apc_dispatcher,
|
|
"movq 0x28(%rsp),%rsi\n\t" /* func */
|
|
"movq 0x30(%rsp),%rdi\n\t" /* dispatcher */
|
|
"movq %gs:0x30,%rbx\n\t"
|
|
"jrcxz 1f\n\t"
|
|
"movq 0x98(%rcx),%rax\n\t" /* context_ptr->Rsp */
|
|
"leaq -0x5c0(%rax),%rsp\n\t" /* sizeof(CONTEXT) + offsetof(frame,ret_addr) */
|
|
"andq $~15,%rsp\n\t"
|
|
"jmp 2f\n"
|
|
"1:\tmovq 0x328(%rbx),%rax\n\t" /* amd64_thread_data()->syscall_frame */
|
|
"leaq -0x4d0(%rax),%rsp\n\t"
|
|
"andq $~15,%rsp\n\t"
|
|
"movq %rdx,%r12\n\t" /* ctx */
|
|
"movq %r8,%r13\n\t" /* arg1 */
|
|
"movq %r9,%r14\n\t" /* arg2 */
|
|
"movq %rsp,%rdx\n\t" /* context */
|
|
"movl $0x10000b,0x30(%rdx)\n\t" /* context.ContextFlags */
|
|
"movq $~1,%rcx\n\t"
|
|
"call " __ASM_NAME("NtGetContextThread") "\n\t"
|
|
"movq %rsp,%rcx\n\t" /* context */
|
|
"movq $0xc0,0x78(%rcx)\n\t" /* context.Rax = STATUS_USER_APC */
|
|
"movq %r12,%rdx\n\t" /* ctx */
|
|
"movq %r13,%r8\n\t" /* arg1 */
|
|
"movq %r14,%r9\n" /* arg2 */
|
|
"2:\tmovq $0,0x328(%rbx)\n\t"
|
|
"movq %rsi,0x20(%rsp)\n\t" /* func */
|
|
"movq 0xa0(%rcx),%rbp\n\t" /* context.Rbp */
|
|
"pushq 0xf8(%rcx)\n\t" /* context.Rip */
|
|
"jmp *%rdi" )
|
|
|
|
|
|
/***********************************************************************
|
|
* call_raise_user_exception_dispatcher
|
|
*/
|
|
__ASM_GLOBAL_FUNC( call_raise_user_exception_dispatcher,
|
|
"movq %gs:0x30,%rdx\n\t"
|
|
"movq 0x328(%rdx),%rax\n\t" /* amd64_thread_data()->syscall_frame */
|
|
"movdqu 0x0(%rax),%xmm6\n\t" /* frame->xmm[0..19] */
|
|
"movdqu 0x10(%rax),%xmm7\n\t"
|
|
"movdqu 0x20(%rax),%xmm8\n\t"
|
|
"movdqu 0x30(%rax),%xmm9\n\t"
|
|
"movdqu 0x40(%rax),%xmm10\n\t"
|
|
"movdqu 0x50(%rax),%xmm11\n\t"
|
|
"movdqu 0x60(%rax),%xmm12\n\t"
|
|
"movdqu 0x70(%rax),%xmm13\n\t"
|
|
"movdqu 0x80(%rax),%xmm14\n\t"
|
|
"movdqu 0x90(%rax),%xmm15\n\t"
|
|
"ldmxcsr 0xa0(%rax)\n\t" /* frame->mxcsr */
|
|
"movq 0xa8(%rax),%r12\n\t" /* frame->r12 */
|
|
"movq 0xb0(%rax),%r13\n\t" /* frame->r13 */
|
|
"movq 0xb8(%rax),%r14\n\t" /* frame->r14 */
|
|
"movq 0xc0(%rax),%r15\n\t" /* frame->r15 */
|
|
"movq 0xc8(%rax),%rdi\n\t" /* frame->rdi */
|
|
"movq 0xd0(%rax),%rsi\n\t" /* frame->rsi */
|
|
"movq 0xd8(%rax),%rbx\n\t" /* frame->rbx */
|
|
"movq 0xe0(%rax),%rbp\n\t" /* frame->rbp */
|
|
"movq $0,0x328(%rdx)\n\t"
|
|
"leaq 0xf0(%rax),%rsp\n\t"
|
|
"jmpq *%rcx" )
|
|
|
|
|
|
/***********************************************************************
|
|
* call_user_exception_dispatcher
|
|
*/
|
|
|
|
extern void WINAPI user_exception_dispatcher_trampoline( struct stack_layout *stack,
|
|
void *pKiUserExceptionDispatcher );
|
|
|
|
__ASM_GLOBAL_FUNC( user_exception_dispatcher_trampoline,
|
|
"movq %rcx,%rsp\n\t"
|
|
"movq 0x98(%rsp),%rcx\n\t" /* context->Rsp */
|
|
"movq 0xa0(%rsp),%rbp\n\t"
|
|
"movq 0xa8(%rsp),%rsi\n\t"
|
|
"movq 0xb0(%rsp),%rdi\n\t"
|
|
"jmpq *%rdx")
|
|
|
|
void WINAPI do_call_user_exception_dispatcher( EXCEPTION_RECORD *rec, CONTEXT *context,
|
|
NTSTATUS (WINAPI *dispatcher)(EXCEPTION_RECORD*,CONTEXT*),
|
|
struct stack_layout *stack )
|
|
{
|
|
if ((context->ContextFlags & CONTEXT_XSTATE) == CONTEXT_XSTATE)
|
|
{
|
|
CONTEXT_EX *xctx = (CONTEXT_EX *)context + 1;
|
|
XSTATE *xs, *src_xs, xs_buf;
|
|
|
|
src_xs = xstate_from_context(context);
|
|
if ((CONTEXT *)src_xs >= &stack->context + 1 || src_xs + 1 <= (XSTATE *)&stack->context)
|
|
{
|
|
xs = src_xs;
|
|
}
|
|
else
|
|
{
|
|
xs = &xs_buf;
|
|
memcpy(xs, src_xs, sizeof(*xs));
|
|
}
|
|
|
|
memmove(&stack->context, context, sizeof(*context) + sizeof(*xctx));
|
|
assert(!((ULONG_PTR)stack->xstate & 63));
|
|
context_init_xstate(&stack->context, stack->xstate);
|
|
memcpy(stack->xstate, xs, sizeof(*xs));
|
|
}
|
|
else
|
|
{
|
|
memmove(&stack->context, context, sizeof(*context));
|
|
}
|
|
memcpy(&stack->rec, rec, sizeof(*rec));
|
|
|
|
/* fix up instruction pointer in context for EXCEPTION_BREAKPOINT */
|
|
if (stack->rec.ExceptionCode == EXCEPTION_BREAKPOINT) stack->context.Rip--;
|
|
|
|
amd64_thread_data()->syscall_frame = NULL;
|
|
user_exception_dispatcher_trampoline( stack, dispatcher );
|
|
}
|
|
|
|
__ASM_GLOBAL_FUNC( call_user_exception_dispatcher,
|
|
"movq 0x98(%rdx),%r9\n\t" /* context->Rsp */
|
|
"andq $~0xf,%r9\n\t"
|
|
"btl $6,0x30(%rdx)\n\t" /* context->ContextFlags, CONTEXT_XSTATE bit. */
|
|
"jnc 1f\n\t"
|
|
"subq $0x140,%r9\n\t" /* sizeof(XSTATE) */
|
|
"andq $~63,%r9\n"
|
|
"1:\tsubq $0x5b0,%r9\n\t" /* sizeof(struct stack_layout) */
|
|
"cmpq %rsp,%r9\n\t"
|
|
"cmovbq %r9,%rsp\n\t"
|
|
"jmp " __ASM_NAME("do_call_user_exception_dispatcher") "\n\t")
|
|
|
|
/***********************************************************************
|
|
* is_privileged_instr
|
|
*
|
|
* Check if the fault location is a privileged instruction.
|
|
*/
|
|
static inline DWORD is_privileged_instr( CONTEXT *context )
|
|
{
|
|
BYTE instr[16];
|
|
unsigned int i, prefix_count = 0;
|
|
unsigned int len = virtual_uninterrupted_read_memory( (BYTE *)context->Rip, instr, sizeof(instr) );
|
|
|
|
for (i = 0; i < len; i++) switch (instr[i])
|
|
{
|
|
/* instruction prefixes */
|
|
case 0x2e: /* %cs: */
|
|
case 0x36: /* %ss: */
|
|
case 0x3e: /* %ds: */
|
|
case 0x26: /* %es: */
|
|
case 0x40: /* rex */
|
|
case 0x41: /* rex */
|
|
case 0x42: /* rex */
|
|
case 0x43: /* rex */
|
|
case 0x44: /* rex */
|
|
case 0x45: /* rex */
|
|
case 0x46: /* rex */
|
|
case 0x47: /* rex */
|
|
case 0x48: /* rex */
|
|
case 0x49: /* rex */
|
|
case 0x4a: /* rex */
|
|
case 0x4b: /* rex */
|
|
case 0x4c: /* rex */
|
|
case 0x4d: /* rex */
|
|
case 0x4e: /* rex */
|
|
case 0x4f: /* rex */
|
|
case 0x64: /* %fs: */
|
|
case 0x65: /* %gs: */
|
|
case 0x66: /* opcode size */
|
|
case 0x67: /* addr size */
|
|
case 0xf0: /* lock */
|
|
case 0xf2: /* repne */
|
|
case 0xf3: /* repe */
|
|
if (++prefix_count >= 15) return EXCEPTION_ILLEGAL_INSTRUCTION;
|
|
continue;
|
|
|
|
case 0x0f: /* extended instruction */
|
|
if (i == len - 1) return 0;
|
|
switch (instr[i + 1])
|
|
{
|
|
case 0x06: /* clts */
|
|
case 0x08: /* invd */
|
|
case 0x09: /* wbinvd */
|
|
case 0x20: /* mov crX, reg */
|
|
case 0x21: /* mov drX, reg */
|
|
case 0x22: /* mov reg, crX */
|
|
case 0x23: /* mov reg drX */
|
|
return EXCEPTION_PRIV_INSTRUCTION;
|
|
}
|
|
return 0;
|
|
case 0x6c: /* insb (%dx) */
|
|
case 0x6d: /* insl (%dx) */
|
|
case 0x6e: /* outsb (%dx) */
|
|
case 0x6f: /* outsl (%dx) */
|
|
case 0xcd: /* int $xx */
|
|
case 0xe4: /* inb al,XX */
|
|
case 0xe5: /* in (e)ax,XX */
|
|
case 0xe6: /* outb XX,al */
|
|
case 0xe7: /* out XX,(e)ax */
|
|
case 0xec: /* inb (%dx),%al */
|
|
case 0xed: /* inl (%dx),%eax */
|
|
case 0xee: /* outb %al,(%dx) */
|
|
case 0xef: /* outl %eax,(%dx) */
|
|
case 0xf4: /* hlt */
|
|
case 0xfa: /* cli */
|
|
case 0xfb: /* sti */
|
|
return EXCEPTION_PRIV_INSTRUCTION;
|
|
default:
|
|
return 0;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* handle_interrupt
|
|
*
|
|
* Handle an interrupt.
|
|
*/
|
|
static inline BOOL handle_interrupt( ucontext_t *sigcontext, EXCEPTION_RECORD *rec, struct xcontext *xcontext )
|
|
{
|
|
CONTEXT *context = &xcontext->c;
|
|
|
|
switch (ERROR_sig(sigcontext) >> 3)
|
|
{
|
|
case 0x2c:
|
|
rec->ExceptionCode = STATUS_ASSERTION_FAILURE;
|
|
break;
|
|
case 0x2d:
|
|
switch (context->Rax)
|
|
{
|
|
case 1: /* BREAKPOINT_PRINT */
|
|
case 3: /* BREAKPOINT_LOAD_SYMBOLS */
|
|
case 4: /* BREAKPOINT_UNLOAD_SYMBOLS */
|
|
case 5: /* BREAKPOINT_COMMAND_STRING (>= Win2003) */
|
|
RIP_sig(sigcontext) += 3;
|
|
return TRUE;
|
|
}
|
|
context->Rip += 3;
|
|
rec->ExceptionCode = EXCEPTION_BREAKPOINT;
|
|
rec->ExceptionAddress = (void *)context->Rip;
|
|
rec->NumberParameters = 1;
|
|
rec->ExceptionInformation[0] = context->Rax;
|
|
break;
|
|
default:
|
|
return FALSE;
|
|
}
|
|
setup_raise_exception( sigcontext, rec, xcontext );
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* handle_syscall_fault
|
|
*
|
|
* Handle a page fault happening during a system call.
|
|
*/
|
|
static BOOL handle_syscall_fault( ucontext_t *sigcontext, EXCEPTION_RECORD *rec, CONTEXT *context )
|
|
{
|
|
struct syscall_frame *frame = amd64_thread_data()->syscall_frame;
|
|
__WINE_FRAME *wine_frame = (__WINE_FRAME *)NtCurrentTeb()->Tib.ExceptionList;
|
|
DWORD i;
|
|
|
|
if (!frame) return FALSE;
|
|
|
|
TRACE( "code=%x flags=%x addr=%p ip=%lx tid=%04x\n",
|
|
rec->ExceptionCode, rec->ExceptionFlags, rec->ExceptionAddress,
|
|
context->Rip, GetCurrentThreadId() );
|
|
for (i = 0; i < rec->NumberParameters; i++)
|
|
TRACE( " info[%d]=%016lx\n", i, rec->ExceptionInformation[i] );
|
|
TRACE(" rax=%016lx rbx=%016lx rcx=%016lx rdx=%016lx\n",
|
|
context->Rax, context->Rbx, context->Rcx, context->Rdx );
|
|
TRACE(" rsi=%016lx rdi=%016lx rbp=%016lx rsp=%016lx\n",
|
|
context->Rsi, context->Rdi, context->Rbp, context->Rsp );
|
|
TRACE(" r8=%016lx r9=%016lx r10=%016lx r11=%016lx\n",
|
|
context->R8, context->R9, context->R10, context->R11 );
|
|
TRACE(" r12=%016lx r13=%016lx r14=%016lx r15=%016lx\n",
|
|
context->R12, context->R13, context->R14, context->R15 );
|
|
|
|
if ((char *)wine_frame < (char *)frame)
|
|
{
|
|
TRACE( "returning to handler\n" );
|
|
RCX_sig(sigcontext) = (ULONG_PTR)&wine_frame->jmp;
|
|
RDX_sig(sigcontext) = 1;
|
|
RIP_sig(sigcontext) = (ULONG_PTR)__wine_longjmp;
|
|
}
|
|
else
|
|
{
|
|
XMM_SAVE_AREA32 *fpu = FPU_sig(sigcontext);
|
|
|
|
TRACE( "returning to user mode ip=%016lx ret=%08x\n", frame->ret_addr, rec->ExceptionCode );
|
|
RAX_sig(sigcontext) = rec->ExceptionCode;
|
|
RBX_sig(sigcontext) = frame->rbx;
|
|
RSI_sig(sigcontext) = frame->rsi;
|
|
RDI_sig(sigcontext) = frame->rdi;
|
|
RBP_sig(sigcontext) = frame->rbp;
|
|
R12_sig(sigcontext) = frame->r12;
|
|
R13_sig(sigcontext) = frame->r13;
|
|
R14_sig(sigcontext) = frame->r14;
|
|
R15_sig(sigcontext) = frame->r15;
|
|
RSP_sig(sigcontext) = (ULONG_PTR)&frame->ret_addr;
|
|
RIP_sig(sigcontext) = frame->thunk_addr;
|
|
if (fpu)
|
|
{
|
|
fpu->MxCsr =frame->mxcsr;
|
|
memcpy( fpu->XmmRegisters + 6, frame->xmm, sizeof(frame->xmm) );
|
|
}
|
|
amd64_thread_data()->syscall_frame = NULL;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* segv_handler
|
|
*
|
|
* Handler for SIGSEGV and related errors.
|
|
*/
|
|
static void segv_handler( int signal, siginfo_t *siginfo, void *sigcontext )
|
|
{
|
|
EXCEPTION_RECORD rec = { 0 };
|
|
struct xcontext context;
|
|
ucontext_t *ucontext = sigcontext;
|
|
|
|
rec.ExceptionAddress = (void *)RIP_sig(ucontext);
|
|
save_context( &context, sigcontext );
|
|
|
|
switch(TRAP_sig(ucontext))
|
|
{
|
|
case TRAP_x86_OFLOW: /* Overflow exception */
|
|
rec.ExceptionCode = EXCEPTION_INT_OVERFLOW;
|
|
break;
|
|
case TRAP_x86_BOUND: /* Bound range exception */
|
|
rec.ExceptionCode = EXCEPTION_ARRAY_BOUNDS_EXCEEDED;
|
|
break;
|
|
case TRAP_x86_PRIVINFLT: /* Invalid opcode exception */
|
|
rec.ExceptionCode = EXCEPTION_ILLEGAL_INSTRUCTION;
|
|
break;
|
|
case TRAP_x86_STKFLT: /* Stack fault */
|
|
rec.ExceptionCode = EXCEPTION_STACK_OVERFLOW;
|
|
break;
|
|
case TRAP_x86_SEGNPFLT: /* Segment not present exception */
|
|
case TRAP_x86_PROTFLT: /* General protection fault */
|
|
{
|
|
WORD err = ERROR_sig(ucontext);
|
|
if (!err && (rec.ExceptionCode = is_privileged_instr( &context.c ))) break;
|
|
if ((err & 7) == 2 && handle_interrupt( ucontext, &rec, &context )) return;
|
|
rec.ExceptionCode = EXCEPTION_ACCESS_VIOLATION;
|
|
rec.NumberParameters = 2;
|
|
rec.ExceptionInformation[0] = 0;
|
|
rec.ExceptionInformation[1] = 0xffffffffffffffff;
|
|
}
|
|
break;
|
|
case TRAP_x86_PAGEFLT: /* Page fault */
|
|
rec.NumberParameters = 2;
|
|
rec.ExceptionInformation[0] = (ERROR_sig(ucontext) >> 1) & 0x09;
|
|
rec.ExceptionInformation[1] = (ULONG_PTR)siginfo->si_addr;
|
|
rec.ExceptionCode = virtual_handle_fault( siginfo->si_addr, rec.ExceptionInformation[0],
|
|
(void *)RSP_sig(ucontext) );
|
|
if (!rec.ExceptionCode) return;
|
|
break;
|
|
case TRAP_x86_ALIGNFLT: /* Alignment check exception */
|
|
rec.ExceptionCode = EXCEPTION_DATATYPE_MISALIGNMENT;
|
|
break;
|
|
default:
|
|
ERR( "Got unexpected trap %ld\n", (ULONG_PTR)TRAP_sig(ucontext) );
|
|
/* fall through */
|
|
case TRAP_x86_NMI: /* NMI interrupt */
|
|
case TRAP_x86_DNA: /* Device not available exception */
|
|
case TRAP_x86_DOUBLEFLT: /* Double fault exception */
|
|
case TRAP_x86_TSSFLT: /* Invalid TSS exception */
|
|
case TRAP_x86_MCHK: /* Machine check exception */
|
|
case TRAP_x86_CACHEFLT: /* Cache flush exception */
|
|
rec.ExceptionCode = EXCEPTION_ILLEGAL_INSTRUCTION;
|
|
break;
|
|
}
|
|
if (handle_syscall_fault( sigcontext, &rec, &context.c )) return;
|
|
setup_raise_exception( sigcontext, &rec, &context );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* trap_handler
|
|
*
|
|
* Handler for SIGTRAP.
|
|
*/
|
|
static void trap_handler( int signal, siginfo_t *siginfo, void *sigcontext )
|
|
{
|
|
EXCEPTION_RECORD rec = { 0 };
|
|
struct xcontext context;
|
|
ucontext_t *ucontext = sigcontext;
|
|
|
|
rec.ExceptionAddress = (void *)RIP_sig(ucontext);
|
|
save_context( &context, sigcontext );
|
|
|
|
switch (siginfo->si_code)
|
|
{
|
|
case TRAP_TRACE: /* Single-step exception */
|
|
case 4 /* TRAP_HWBKPT */: /* Hardware breakpoint exception */
|
|
rec.ExceptionCode = EXCEPTION_SINGLE_STEP;
|
|
break;
|
|
case TRAP_BRKPT: /* Breakpoint exception */
|
|
#ifdef SI_KERNEL
|
|
case SI_KERNEL:
|
|
#endif
|
|
/* Check if this is actually icebp instruction */
|
|
if (((unsigned char *)RIP_sig(ucontext))[-1] == 0xF1)
|
|
{
|
|
rec.ExceptionCode = EXCEPTION_SINGLE_STEP;
|
|
break;
|
|
}
|
|
rec.ExceptionAddress = (char *)rec.ExceptionAddress - 1; /* back up over the int3 instruction */
|
|
/* fall through */
|
|
default:
|
|
rec.ExceptionCode = EXCEPTION_BREAKPOINT;
|
|
rec.NumberParameters = 1;
|
|
rec.ExceptionInformation[0] = 0;
|
|
break;
|
|
}
|
|
setup_raise_exception( sigcontext, &rec, &context );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* fpe_handler
|
|
*
|
|
* Handler for SIGFPE.
|
|
*/
|
|
static void fpe_handler( int signal, siginfo_t *siginfo, void *sigcontext )
|
|
{
|
|
EXCEPTION_RECORD rec = { 0 };
|
|
|
|
switch (siginfo->si_code)
|
|
{
|
|
case FPE_FLTSUB:
|
|
rec.ExceptionCode = EXCEPTION_ARRAY_BOUNDS_EXCEEDED;
|
|
break;
|
|
case FPE_INTDIV:
|
|
rec.ExceptionCode = EXCEPTION_INT_DIVIDE_BY_ZERO;
|
|
break;
|
|
case FPE_INTOVF:
|
|
rec.ExceptionCode = EXCEPTION_INT_OVERFLOW;
|
|
break;
|
|
case FPE_FLTDIV:
|
|
rec.ExceptionCode = EXCEPTION_FLT_DIVIDE_BY_ZERO;
|
|
break;
|
|
case FPE_FLTOVF:
|
|
rec.ExceptionCode = EXCEPTION_FLT_OVERFLOW;
|
|
break;
|
|
case FPE_FLTUND:
|
|
rec.ExceptionCode = EXCEPTION_FLT_UNDERFLOW;
|
|
break;
|
|
case FPE_FLTRES:
|
|
rec.ExceptionCode = EXCEPTION_FLT_INEXACT_RESULT;
|
|
break;
|
|
case FPE_FLTINV:
|
|
default:
|
|
rec.ExceptionCode = EXCEPTION_FLT_INVALID_OPERATION;
|
|
break;
|
|
}
|
|
setup_exception( sigcontext, &rec );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* int_handler
|
|
*
|
|
* Handler for SIGINT.
|
|
*/
|
|
static void int_handler( int signal, siginfo_t *siginfo, void *sigcontext )
|
|
{
|
|
EXCEPTION_RECORD rec = { CONTROL_C_EXIT };
|
|
|
|
setup_exception( sigcontext, &rec );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* abrt_handler
|
|
*
|
|
* Handler for SIGABRT.
|
|
*/
|
|
static void abrt_handler( int signal, siginfo_t *siginfo, void *sigcontext )
|
|
{
|
|
EXCEPTION_RECORD rec = { EXCEPTION_WINE_ASSERTION, EH_NONCONTINUABLE };
|
|
|
|
setup_exception( sigcontext, &rec );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* quit_handler
|
|
*
|
|
* Handler for SIGQUIT.
|
|
*/
|
|
static void quit_handler( int signal, siginfo_t *siginfo, void *ucontext )
|
|
{
|
|
abort_thread(0);
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* usr1_handler
|
|
*
|
|
* Handler for SIGUSR1, used to signal a thread that it got suspended.
|
|
*/
|
|
static void usr1_handler( int signal, siginfo_t *siginfo, void *ucontext )
|
|
{
|
|
struct xcontext context;
|
|
|
|
save_context( &context, ucontext );
|
|
wait_suspend( &context.c );
|
|
restore_context( &context, ucontext );
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* get_thread_ldt_entry
|
|
*/
|
|
NTSTATUS get_thread_ldt_entry( HANDLE handle, void *data, ULONG len, ULONG *ret_len )
|
|
{
|
|
return STATUS_NOT_IMPLEMENTED;
|
|
}
|
|
|
|
|
|
/******************************************************************************
|
|
* NtSetLdtEntries (NTDLL.@)
|
|
* ZwSetLdtEntries (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtSetLdtEntries( ULONG sel1, LDT_ENTRY entry1, ULONG sel2, LDT_ENTRY entry2 )
|
|
{
|
|
return STATUS_NOT_IMPLEMENTED;
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* signal_init_threading
|
|
*/
|
|
void signal_init_threading(void)
|
|
{
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* signal_alloc_thread
|
|
*/
|
|
NTSTATUS signal_alloc_thread( TEB *teb )
|
|
{
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* signal_free_thread
|
|
*/
|
|
void signal_free_thread( TEB *teb )
|
|
{
|
|
}
|
|
|
|
#ifdef __APPLE__
|
|
/**********************************************************************
|
|
* mac_thread_gsbase
|
|
*/
|
|
static void *mac_thread_gsbase(void)
|
|
{
|
|
struct thread_identifier_info tiinfo;
|
|
unsigned int info_count = THREAD_IDENTIFIER_INFO_COUNT;
|
|
static int gsbase_offset = -1;
|
|
|
|
kern_return_t kr = thread_info(mach_thread_self(), THREAD_IDENTIFIER_INFO, (thread_info_t) &tiinfo, &info_count);
|
|
if (kr == KERN_SUCCESS) return (void*)tiinfo.thread_handle;
|
|
|
|
if (gsbase_offset < 0)
|
|
{
|
|
/* Search for the array of TLS slots within the pthread data structure.
|
|
That's what the macOS pthread implementation uses for gsbase. */
|
|
const void* const sentinel1 = (const void*)0x2bffb6b4f11228ae;
|
|
const void* const sentinel2 = (const void*)0x0845a7ff6ab76707;
|
|
int rc;
|
|
pthread_key_t key;
|
|
const void** p = (const void**)pthread_self();
|
|
int i;
|
|
|
|
gsbase_offset = 0;
|
|
if ((rc = pthread_key_create(&key, NULL))) return NULL;
|
|
|
|
pthread_setspecific(key, sentinel1);
|
|
|
|
for (i = key + 1; i < 2000; i++) /* arbitrary limit */
|
|
{
|
|
if (p[i] == sentinel1)
|
|
{
|
|
pthread_setspecific(key, sentinel2);
|
|
|
|
if (p[i] == sentinel2)
|
|
{
|
|
gsbase_offset = (i - key) * sizeof(*p);
|
|
break;
|
|
}
|
|
|
|
pthread_setspecific(key, sentinel1);
|
|
}
|
|
}
|
|
|
|
pthread_key_delete(key);
|
|
}
|
|
|
|
if (gsbase_offset) return (char*)pthread_self() + gsbase_offset;
|
|
return NULL;
|
|
}
|
|
#endif
|
|
|
|
|
|
/**********************************************************************
|
|
* signal_init_thread
|
|
*/
|
|
void signal_init_thread( TEB *teb )
|
|
{
|
|
const WORD fpu_cw = 0x27f;
|
|
|
|
#if defined __linux__
|
|
arch_prctl( ARCH_SET_GS, teb );
|
|
#elif defined (__FreeBSD__) || defined (__FreeBSD_kernel__)
|
|
amd64_set_gsbase( teb );
|
|
#elif defined(__NetBSD__)
|
|
sysarch( X86_64_SET_GSBASE, &teb );
|
|
#elif defined (__APPLE__)
|
|
__asm__ volatile (".byte 0x65\n\tmovq %0,%c1"
|
|
:
|
|
: "r" (teb->Tib.Self), "n" (FIELD_OFFSET(TEB, Tib.Self)));
|
|
__asm__ volatile (".byte 0x65\n\tmovq %0,%c1"
|
|
:
|
|
: "r" (teb->ThreadLocalStoragePointer), "n" (FIELD_OFFSET(TEB, ThreadLocalStoragePointer)));
|
|
|
|
/* alloc_tls_slot() needs to poke a value to an address relative to each
|
|
thread's gsbase. Have each thread record its gsbase pointer into its
|
|
TEB so alloc_tls_slot() can find it. */
|
|
teb->Reserved5[0] = mac_thread_gsbase();
|
|
#else
|
|
# error Please define setting %gs for your architecture
|
|
#endif
|
|
|
|
#ifdef __GNUC__
|
|
__asm__ volatile ("fninit; fldcw %0" : : "m" (fpu_cw));
|
|
#else
|
|
FIXME("FPU setup not implemented for this platform.\n");
|
|
#endif
|
|
}
|
|
|
|
|
|
/**********************************************************************
|
|
* signal_init_process
|
|
*/
|
|
void signal_init_process(void)
|
|
{
|
|
struct sigaction sig_act;
|
|
|
|
sig_act.sa_mask = server_block_set;
|
|
sig_act.sa_flags = SA_SIGINFO | SA_RESTART | SA_ONSTACK;
|
|
|
|
sig_act.sa_sigaction = int_handler;
|
|
if (sigaction( SIGINT, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = fpe_handler;
|
|
if (sigaction( SIGFPE, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = abrt_handler;
|
|
if (sigaction( SIGABRT, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = quit_handler;
|
|
if (sigaction( SIGQUIT, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = usr1_handler;
|
|
if (sigaction( SIGUSR1, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = trap_handler;
|
|
if (sigaction( SIGTRAP, &sig_act, NULL ) == -1) goto error;
|
|
sig_act.sa_sigaction = segv_handler;
|
|
if (sigaction( SIGSEGV, &sig_act, NULL ) == -1) goto error;
|
|
if (sigaction( SIGILL, &sig_act, NULL ) == -1) goto error;
|
|
if (sigaction( SIGBUS, &sig_act, NULL ) == -1) goto error;
|
|
return;
|
|
|
|
error:
|
|
perror("sigaction");
|
|
exit(1);
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* init_thread_context
|
|
*/
|
|
static void init_thread_context( CONTEXT *context, LPTHREAD_START_ROUTINE entry, void *arg )
|
|
{
|
|
__asm__( "movw %%cs,%0" : "=m" (context->SegCs) );
|
|
__asm__( "movw %%ss,%0" : "=m" (context->SegSs) );
|
|
context->Rcx = (ULONG_PTR)entry;
|
|
context->Rdx = (ULONG_PTR)arg;
|
|
context->Rsp = (ULONG_PTR)NtCurrentTeb()->Tib.StackBase - 0x28;
|
|
context->Rip = (ULONG_PTR)pRtlUserThreadStart;
|
|
context->EFlags = 0x200;
|
|
context->u.FltSave.ControlWord = 0x27f;
|
|
context->u.FltSave.MxCsr = context->MxCsr = 0x1f80;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* get_initial_context
|
|
*/
|
|
PCONTEXT DECLSPEC_HIDDEN get_initial_context( LPTHREAD_START_ROUTINE entry, void *arg, BOOL suspend )
|
|
{
|
|
CONTEXT *ctx;
|
|
|
|
if (suspend)
|
|
{
|
|
CONTEXT context = { 0 };
|
|
|
|
context.ContextFlags = CONTEXT_ALL;
|
|
init_thread_context( &context, entry, arg );
|
|
wait_suspend( &context );
|
|
ctx = (CONTEXT *)((ULONG_PTR)context.Rsp & ~15) - 1;
|
|
*ctx = context;
|
|
}
|
|
else
|
|
{
|
|
ctx = (CONTEXT *)((char *)NtCurrentTeb()->Tib.StackBase - 0x30) - 1;
|
|
init_thread_context( ctx, entry, arg );
|
|
}
|
|
pthread_sigmask( SIG_UNBLOCK, &server_block_set, NULL );
|
|
ctx->ContextFlags = CONTEXT_FULL;
|
|
return ctx;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* signal_start_thread
|
|
*/
|
|
__ASM_GLOBAL_FUNC( signal_start_thread,
|
|
"subq $56,%rsp\n\t"
|
|
__ASM_SEH(".seh_stackalloc 56\n\t")
|
|
__ASM_SEH(".seh_endprologue\n\t")
|
|
__ASM_CFI(".cfi_adjust_cfa_offset 56\n\t")
|
|
"movq %rbp,48(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %rbp,48\n\t")
|
|
"movq %rbx,40(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %rbx,40\n\t")
|
|
"movq %r12,32(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %r12,32\n\t")
|
|
"movq %r13,24(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %r13,24\n\t")
|
|
"movq %r14,16(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %r14,16\n\t")
|
|
"movq %r15,8(%rsp)\n\t"
|
|
__ASM_CFI(".cfi_rel_offset %r15,8\n\t")
|
|
/* store exit frame */
|
|
"movq %gs:0x30,%rax\n\t"
|
|
"movq %rsp,0x320(%rax)\n\t" /* amd64_thread_data()->exit_frame */
|
|
/* switch to thread stack */
|
|
"movq 8(%rax),%rax\n\t" /* NtCurrentTeb()->Tib.StackBase */
|
|
"movq %rcx,%rbx\n\t" /* thunk */
|
|
"leaq -0x1000(%rax),%rsp\n\t"
|
|
/* attach dlls */
|
|
"call " __ASM_NAME("get_initial_context") "\n\t"
|
|
"movq %rax,%rcx\n\t" /* context */
|
|
"xorq %rax,%rax\n\t"
|
|
"pushq %rax\n\t"
|
|
"jmp *%rbx" )
|
|
|
|
|
|
/***********************************************************************
|
|
* signal_exit_thread
|
|
*/
|
|
__ASM_GLOBAL_FUNC( signal_exit_thread,
|
|
/* fetch exit frame */
|
|
"movq %gs:0x30,%rax\n\t"
|
|
"movq 0x320(%rax),%rdx\n\t" /* amd64_thread_data()->exit_frame */
|
|
"testq %rdx,%rdx\n\t"
|
|
"jnz 1f\n\t"
|
|
"jmp *%rsi\n"
|
|
/* switch to exit frame stack */
|
|
"1:\tmovq $0,0x330(%rax)\n\t"
|
|
"movq %rdx,%rsp\n\t"
|
|
__ASM_CFI(".cfi_adjust_cfa_offset 56\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %rbp,48\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %rbx,40\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %r12,32\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %r13,24\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %r14,16\n\t")
|
|
__ASM_CFI(".cfi_rel_offset %r15,8\n\t")
|
|
"call *%rsi" )
|
|
|
|
#endif /* __x86_64__ */
|