/* * Win32 virtual memory functions * * Copyright 1997, 2002, 2020 Alexandre Julliard * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA */ #if 0 #pragma makedep unix #endif #include "config.h" #include "wine/port.h" #include #include #include #include #include #include #ifdef HAVE_SYS_SOCKET_H # include #endif #ifdef HAVE_SYS_STAT_H # include #endif #ifdef HAVE_SYS_MMAN_H # include #endif #ifdef HAVE_SYS_SYSINFO_H # include #endif #ifdef HAVE_VALGRIND_VALGRIND_H # include #endif #if defined(__APPLE__) # include # include #endif #include "ntstatus.h" #define WIN32_NO_STATUS #include "windef.h" #include "winnt.h" #include "winternl.h" #include "wine/library.h" #include "wine/exception.h" #include "wine/list.h" #include "wine/rbtree.h" #include "unix_private.h" #include "wine/debug.h" WINE_DEFAULT_DEBUG_CHANNEL(virtual); WINE_DECLARE_DEBUG_CHANNEL(module); struct preload_info { void *addr; size_t size; }; struct reserved_area { struct list entry; void *base; size_t size; }; static struct list reserved_areas = LIST_INIT(reserved_areas); struct file_view { struct wine_rb_entry entry; /* entry in global view tree */ void *base; /* base address */ size_t size; /* size in bytes */ unsigned int protect; /* protection for all pages at allocation time and SEC_* flags */ }; /* per-page protection flags */ #define VPROT_READ 0x01 #define VPROT_WRITE 0x02 #define VPROT_EXEC 0x04 #define VPROT_WRITECOPY 0x08 #define VPROT_GUARD 0x10 #define VPROT_COMMITTED 0x20 #define VPROT_WRITEWATCH 0x40 /* per-mapping protection flags */ #define VPROT_SYSTEM 0x0200 /* system view (underlying mmap not under our control) */ /* Conversion from VPROT_* to Win32 flags */ static const BYTE VIRTUAL_Win32Flags[16] = { PAGE_NOACCESS, /* 0 */ PAGE_READONLY, /* READ */ PAGE_READWRITE, /* WRITE */ PAGE_READWRITE, /* READ | WRITE */ PAGE_EXECUTE, /* EXEC */ PAGE_EXECUTE_READ, /* READ | EXEC */ PAGE_EXECUTE_READWRITE, /* WRITE | EXEC */ PAGE_EXECUTE_READWRITE, /* READ | WRITE | EXEC */ PAGE_WRITECOPY, /* WRITECOPY */ PAGE_WRITECOPY, /* READ | WRITECOPY */ PAGE_WRITECOPY, /* WRITE | WRITECOPY */ PAGE_WRITECOPY, /* READ | WRITE | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* READ | EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* WRITE | EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY /* READ | WRITE | EXEC | WRITECOPY */ }; static struct wine_rb_tree views_tree; static pthread_mutex_t virtual_mutex; static const BOOL is_win64 = (sizeof(void *) > sizeof(int)); static const UINT page_shift = 12; static const UINT_PTR page_mask = 0xfff; static const UINT_PTR granularity_mask = 0xffff; /* Note: these are Windows limits, you cannot change them. */ #ifdef __i386__ static void *address_space_start = (void *)0x110000; /* keep DOS area clear */ #else static void *address_space_start = (void *)0x10000; #endif #ifdef __aarch64__ static void *address_space_limit = (void *)0xffffffff0000; /* top of the total available address space */ #elif defined(_WIN64) static void *address_space_limit = (void *)0x7fffffff0000; #else static void *address_space_limit = (void *)0xc0000000; #endif #ifdef _WIN64 static void *user_space_limit = (void *)0x7fffffff0000; /* top of the user address space */ static void *working_set_limit = (void *)0x7fffffff0000; /* top of the current working set */ #else static void *user_space_limit = (void *)0x7fff0000; static void *working_set_limit = (void *)0x7fff0000; #endif struct _KUSER_SHARED_DATA *user_shared_data = (void *)0x7ffe0000; SIZE_T signal_stack_size = 0; SIZE_T signal_stack_mask = 0; static SIZE_T signal_stack_align; /* TEB allocation blocks */ static TEB *teb_block; static TEB *next_free_teb; static int teb_block_pos; static struct list teb_list = LIST_INIT( teb_list ); #define ROUND_ADDR(addr,mask) ((void *)((UINT_PTR)(addr) & ~(UINT_PTR)(mask))) #define ROUND_SIZE(addr,size) (((SIZE_T)(size) + ((UINT_PTR)(addr) & page_mask) + page_mask) & ~page_mask) #define VIRTUAL_DEBUG_DUMP_VIEW(view) do { if (TRACE_ON(virtual)) dump_view(view); } while (0) #ifndef MAP_NORESERVE #define MAP_NORESERVE 0 #endif #ifndef MAP_TRYFIXED #define MAP_TRYFIXED 0 #endif #ifdef _WIN64 /* on 64-bit the page protection bytes use a 2-level table */ static const size_t pages_vprot_shift = 20; static const size_t pages_vprot_mask = (1 << 20) - 1; static size_t pages_vprot_size; static BYTE **pages_vprot; #else /* on 32-bit we use a simple array with one byte per page */ static BYTE *pages_vprot; #endif static struct file_view *view_block_start, *view_block_end, *next_free_view; static const size_t view_block_size = 0x100000; static void *preload_reserve_start; static void *preload_reserve_end; static BOOL force_exec_prot; /* whether to force PROT_EXEC on all PROT_READ mmaps */ struct range_entry { void *base; void *end; }; static struct range_entry *free_ranges; static struct range_entry *free_ranges_end; static void reserve_area( void *addr, void *end ) { #ifdef __APPLE__ #ifdef __i386__ static const mach_vm_address_t max_address = VM_MAX_ADDRESS; #else static const mach_vm_address_t max_address = MACH_VM_MAX_ADDRESS; #endif mach_vm_address_t address = (mach_vm_address_t)addr; mach_vm_address_t end_address = (mach_vm_address_t)end; if (!end_address || max_address < end_address) end_address = max_address; while (address < end_address) { mach_vm_address_t hole_address = address; kern_return_t ret; mach_vm_size_t size; vm_region_basic_info_data_64_t info; mach_msg_type_number_t count = VM_REGION_BASIC_INFO_COUNT_64; mach_port_t dummy_object_name = MACH_PORT_NULL; /* find the mapped region at or above the current address. */ ret = mach_vm_region(mach_task_self(), &address, &size, VM_REGION_BASIC_INFO_64, (vm_region_info_t)&info, &count, &dummy_object_name); if (ret != KERN_SUCCESS) { address = max_address; size = 0; } if (end_address < address) address = end_address; if (hole_address < address) { /* found a hole, attempt to reserve it. */ size_t hole_size = address - hole_address; mach_vm_address_t alloc_address = hole_address; ret = mach_vm_map( mach_task_self(), &alloc_address, hole_size, 0, VM_FLAGS_FIXED, MEMORY_OBJECT_NULL, 0, 0, PROT_NONE, VM_PROT_ALL, VM_INHERIT_COPY ); if (!ret) mmap_add_reserved_area( (void*)hole_address, hole_size ); else if (ret == KERN_NO_SPACE) { /* something filled (part of) the hole before we could. go back and look again. */ address = hole_address; continue; } } address += size; } #else void *ptr; int flags = MAP_PRIVATE | MAP_ANON | MAP_NORESERVE | MAP_TRYFIXED; size_t size = (char *)end - (char *)addr; if (!size) return; #if defined(__FreeBSD__) || defined(__FreeBSD_kernel__) ptr = mmap( addr, size, PROT_NONE, flags | MAP_FIXED | MAP_EXCL, -1, 0 ); #else ptr = mmap( addr, size, PROT_NONE, flags, -1, 0 ); #endif if (ptr == addr) { mmap_add_reserved_area( addr, size ); return; } if (ptr != (void *)-1) munmap( ptr, size ); size = (size / 2) & ~granularity_mask; if (size) { reserve_area( addr, (char *)addr + size ); reserve_area( (char *)addr + size, end ); } #endif /* __APPLE__ */ } static void mmap_init( const struct preload_info *preload_info ) { #ifndef _WIN64 #ifndef __APPLE__ char stack; char * const stack_ptr = &stack; #endif char *user_space_limit = (char *)0x7ffe0000; int i; if (preload_info) { /* check for a reserved area starting at the user space limit */ /* to avoid wasting time trying to allocate it again */ for (i = 0; preload_info[i].size; i++) { if ((char *)preload_info[i].addr > user_space_limit) break; if ((char *)preload_info[i].addr + preload_info[i].size > user_space_limit) { user_space_limit = (char *)preload_info[i].addr + preload_info[i].size; break; } } } else reserve_area( (void *)0x00010000, (void *)0x40000000 ); #ifndef __APPLE__ if (stack_ptr >= user_space_limit) { char *end = 0; char *base = stack_ptr - ((unsigned int)stack_ptr & granularity_mask) - (granularity_mask + 1); if (base > user_space_limit) reserve_area( user_space_limit, base ); base = stack_ptr - ((unsigned int)stack_ptr & granularity_mask) + (granularity_mask + 1); #if defined(linux) || defined(__FreeBSD__) || defined (__FreeBSD_kernel__) || defined(__DragonFly__) /* Heuristic: assume the stack is near the end of the address */ /* space, this avoids a lot of futile allocation attempts */ end = (char *)(((unsigned long)base + 0x0fffffff) & 0xf0000000); #endif reserve_area( base, end ); } else #endif reserve_area( user_space_limit, 0 ); #else if (preload_info) return; /* if we don't have a preloader, try to reserve the space now */ reserve_area( (void *)0x000000010000, (void *)0x000068000000 ); reserve_area( (void *)0x00007ff00000, (void *)0x00007fff0000 ); reserve_area( (void *)0x7ffffe000000, (void *)0x7fffffff0000 ); #endif } void CDECL mmap_add_reserved_area( void *addr, SIZE_T size ) { struct reserved_area *area; struct list *ptr; if (!((char *)addr + size)) size--; /* avoid wrap-around */ LIST_FOR_EACH( ptr, &reserved_areas ) { area = LIST_ENTRY( ptr, struct reserved_area, entry ); if (area->base > addr) { /* try to merge with the next one */ if ((char *)addr + size == (char *)area->base) { area->base = addr; area->size += size; return; } break; } else if ((char *)area->base + area->size == (char *)addr) { /* merge with the previous one */ area->size += size; /* try to merge with the next one too */ if ((ptr = list_next( &reserved_areas, ptr ))) { struct reserved_area *next = LIST_ENTRY( ptr, struct reserved_area, entry ); if ((char *)addr + size == (char *)next->base) { area->size += next->size; list_remove( &next->entry ); free( next ); } } return; } } if ((area = malloc( sizeof(*area) ))) { area->base = addr; area->size = size; list_add_before( ptr, &area->entry ); } } void CDECL mmap_remove_reserved_area( void *addr, SIZE_T size ) { struct reserved_area *area; struct list *ptr; if (!((char *)addr + size)) size--; /* avoid wrap-around */ ptr = list_head( &reserved_areas ); /* find the first area covering address */ while (ptr) { area = LIST_ENTRY( ptr, struct reserved_area, entry ); if ((char *)area->base >= (char *)addr + size) break; /* outside the range */ if ((char *)area->base + area->size > (char *)addr) /* overlaps range */ { if (area->base >= addr) { if ((char *)area->base + area->size > (char *)addr + size) { /* range overlaps beginning of area only -> shrink area */ area->size -= (char *)addr + size - (char *)area->base; area->base = (char *)addr + size; break; } else { /* range contains the whole area -> remove area completely */ ptr = list_next( &reserved_areas, ptr ); list_remove( &area->entry ); free( area ); continue; } } else { if ((char *)area->base + area->size > (char *)addr + size) { /* range is in the middle of area -> split area in two */ struct reserved_area *new_area = malloc( sizeof(*new_area) ); if (new_area) { new_area->base = (char *)addr + size; new_area->size = (char *)area->base + area->size - (char *)new_area->base; list_add_after( ptr, &new_area->entry ); } else size = (char *)area->base + area->size - (char *)addr; area->size = (char *)addr - (char *)area->base; break; } else { /* range overlaps end of area only -> shrink area */ area->size = (char *)addr - (char *)area->base; } } } ptr = list_next( &reserved_areas, ptr ); } } int CDECL mmap_is_in_reserved_area( void *addr, SIZE_T size ) { struct reserved_area *area; struct list *ptr; LIST_FOR_EACH( ptr, &reserved_areas ) { area = LIST_ENTRY( ptr, struct reserved_area, entry ); if (area->base > addr) break; if ((char *)area->base + area->size <= (char *)addr) continue; /* area must contain block completely */ if ((char *)area->base + area->size < (char *)addr + size) return -1; return 1; } return 0; } int CDECL mmap_enum_reserved_areas( int (CDECL *enum_func)(void *base, SIZE_T size, void *arg), void *arg, int top_down ) { int ret = 0; struct list *ptr; if (top_down) { for (ptr = reserved_areas.prev; ptr != &reserved_areas; ptr = ptr->prev) { struct reserved_area *area = LIST_ENTRY( ptr, struct reserved_area, entry ); if ((ret = enum_func( area->base, area->size, arg ))) break; } } else { for (ptr = reserved_areas.next; ptr != &reserved_areas; ptr = ptr->next) { struct reserved_area *area = LIST_ENTRY( ptr, struct reserved_area, entry ); if ((ret = enum_func( area->base, area->size, arg ))) break; } } return ret; } /*********************************************************************** * free_ranges_lower_bound * * Returns the first range whose end is not less than addr, or end if there's none. */ static struct range_entry *free_ranges_lower_bound( void *addr ) { struct range_entry *begin = free_ranges; struct range_entry *end = free_ranges_end; struct range_entry *mid; while (begin < end) { mid = begin + (end - begin) / 2; if (mid->end < addr) begin = mid + 1; else end = mid; } return begin; } /*********************************************************************** * free_ranges_insert_view * * Updates the free_ranges after a new view has been created. */ static void free_ranges_insert_view( struct file_view *view ) { void *view_base = ROUND_ADDR( view->base, granularity_mask ); void *view_end = ROUND_ADDR( (char *)view->base + view->size + granularity_mask, granularity_mask ); struct range_entry *range = free_ranges_lower_bound( view_base ); struct range_entry *next = range + 1; /* free_ranges initial value is such that the view is either inside range or before another one. */ assert( range != free_ranges_end ); assert( range->end > view_base || next != free_ranges_end ); /* this happens because virtual_alloc_thread_stack shrinks a view, then creates another one on top, * or because AT_ROUND_TO_PAGE was used with NtMapViewOfSection to force 4kB aligned mapping. */ if ((range->end > view_base && range->base >= view_end) || (range->end == view_base && next->base >= view_end)) { /* on Win64, assert that it's correctly aligned so we're not going to be in trouble later */ assert( (!is_win64 && !is_wow64) || view->base == view_base ); WARN( "range %p - %p is already mapped\n", view_base, view_end ); return; } /* this should never happen */ if (range->base > view_base || range->end < view_end) ERR( "range %p - %p is already partially mapped\n", view_base, view_end ); assert( range->base <= view_base && range->end >= view_end ); /* need to split the range in two */ if (range->base < view_base && range->end > view_end) { memmove( next + 1, next, (free_ranges_end - next) * sizeof(struct range_entry) ); free_ranges_end += 1; if ((char *)free_ranges_end - (char *)free_ranges > view_block_size) ERR( "Free range sequence is full, trouble ahead!\n" ); assert( (char *)free_ranges_end - (char *)free_ranges <= view_block_size ); next->base = view_end; next->end = range->end; range->end = view_base; } else { /* otherwise we just have to shrink it */ if (range->base < view_base) range->end = view_base; else range->base = view_end; if (range->base < range->end) return; /* and possibly remove it if it's now empty */ memmove( range, next, (free_ranges_end - next) * sizeof(struct range_entry) ); free_ranges_end -= 1; assert( free_ranges_end - free_ranges > 0 ); } } /*********************************************************************** * free_ranges_remove_view * * Updates the free_ranges after a view has been destroyed. */ static void free_ranges_remove_view( struct file_view *view ) { void *view_base = ROUND_ADDR( view->base, granularity_mask ); void *view_end = ROUND_ADDR( (char *)view->base + view->size + granularity_mask, granularity_mask ); struct range_entry *range = free_ranges_lower_bound( view_base ); struct range_entry *next = range + 1; /* It's possible to use AT_ROUND_TO_PAGE on 32bit with NtMapViewOfSection to force 4kB alignment, * and this breaks our assumptions. Look at the views around to check if the range is still in use. */ #ifndef _WIN64 struct file_view *prev_view = WINE_RB_ENTRY_VALUE( wine_rb_prev( &view->entry ), struct file_view, entry ); struct file_view *next_view = WINE_RB_ENTRY_VALUE( wine_rb_next( &view->entry ), struct file_view, entry ); void *prev_view_base = prev_view ? ROUND_ADDR( prev_view->base, granularity_mask ) : NULL; void *prev_view_end = prev_view ? ROUND_ADDR( (char *)prev_view->base + prev_view->size + granularity_mask, granularity_mask ) : NULL; void *next_view_base = next_view ? ROUND_ADDR( next_view->base, granularity_mask ) : NULL; void *next_view_end = next_view ? ROUND_ADDR( (char *)next_view->base + next_view->size + granularity_mask, granularity_mask ) : NULL; if ((prev_view_base < view_end && prev_view_end > view_base) || (next_view_base < view_end && next_view_end > view_base)) { WARN( "range %p - %p is still mapped\n", view_base, view_end ); return; } #endif /* free_ranges initial value is such that the view is either inside range or before another one. */ assert( range != free_ranges_end ); assert( range->end > view_base || next != free_ranges_end ); /* this should never happen, but we can safely ignore it */ if (range->base <= view_base && range->end >= view_end) { WARN( "range %p - %p is already unmapped\n", view_base, view_end ); return; } /* this should never happen */ if (range->base < view_end && range->end > view_base) ERR( "range %p - %p is already partially unmapped\n", view_base, view_end ); assert( range->end <= view_base || range->base >= view_end ); /* merge with next if possible */ if (range->end == view_base && next->base == view_end) { range->end = next->end; memmove( next, next + 1, (free_ranges_end - next - 1) * sizeof(struct range_entry) ); free_ranges_end -= 1; assert( free_ranges_end - free_ranges > 0 ); } /* or try growing the range */ else if (range->end == view_base) range->end = view_end; else if (range->base == view_end) range->base = view_base; /* otherwise create a new one */ else { memmove( range + 1, range, (free_ranges_end - range) * sizeof(struct range_entry) ); free_ranges_end += 1; if ((char *)free_ranges_end - (char *)free_ranges > view_block_size) ERR( "Free range sequence is full, trouble ahead!\n" ); assert( (char *)free_ranges_end - (char *)free_ranges <= view_block_size ); range->base = view_base; range->end = view_end; } } static inline int is_view_valloc( const struct file_view *view ) { return !(view->protect & (SEC_FILE | SEC_RESERVE | SEC_COMMIT)); } /*********************************************************************** * get_page_vprot * * Return the page protection byte. */ static BYTE get_page_vprot( const void *addr ) { size_t idx = (size_t)addr >> page_shift; #ifdef _WIN64 if ((idx >> pages_vprot_shift) >= pages_vprot_size) return 0; if (!pages_vprot[idx >> pages_vprot_shift]) return 0; return pages_vprot[idx >> pages_vprot_shift][idx & pages_vprot_mask]; #else return pages_vprot[idx]; #endif } /*********************************************************************** * set_page_vprot * * Set a range of page protection bytes. */ static void set_page_vprot( const void *addr, size_t size, BYTE vprot ) { size_t idx = (size_t)addr >> page_shift; size_t end = ((size_t)addr + size + page_mask) >> page_shift; #ifdef _WIN64 while (idx >> pages_vprot_shift != end >> pages_vprot_shift) { size_t dir_size = pages_vprot_mask + 1 - (idx & pages_vprot_mask); memset( pages_vprot[idx >> pages_vprot_shift] + (idx & pages_vprot_mask), vprot, dir_size ); idx += dir_size; } memset( pages_vprot[idx >> pages_vprot_shift] + (idx & pages_vprot_mask), vprot, end - idx ); #else memset( pages_vprot + idx, vprot, end - idx ); #endif } /*********************************************************************** * set_page_vprot_bits * * Set or clear bits in a range of page protection bytes. */ static void set_page_vprot_bits( const void *addr, size_t size, BYTE set, BYTE clear ) { size_t idx = (size_t)addr >> page_shift; size_t end = ((size_t)addr + size + page_mask) >> page_shift; #ifdef _WIN64 for ( ; idx < end; idx++) { BYTE *ptr = pages_vprot[idx >> pages_vprot_shift] + (idx & pages_vprot_mask); *ptr = (*ptr & ~clear) | set; } #else for ( ; idx < end; idx++) pages_vprot[idx] = (pages_vprot[idx] & ~clear) | set; #endif } /*********************************************************************** * alloc_pages_vprot * * Allocate the page protection bytes for a given range. */ static BOOL alloc_pages_vprot( const void *addr, size_t size ) { #ifdef _WIN64 size_t idx = (size_t)addr >> page_shift; size_t end = ((size_t)addr + size + page_mask) >> page_shift; size_t i; void *ptr; assert( end <= pages_vprot_size << pages_vprot_shift ); for (i = idx >> pages_vprot_shift; i < (end + pages_vprot_mask) >> pages_vprot_shift; i++) { if (pages_vprot[i]) continue; if ((ptr = wine_anon_mmap( NULL, pages_vprot_mask + 1, PROT_READ | PROT_WRITE, 0 )) == (void *)-1) return FALSE; pages_vprot[i] = ptr; } #endif return TRUE; } /*********************************************************************** * compare_view * * View comparison function used for the rb tree. */ static int compare_view( const void *addr, const struct wine_rb_entry *entry ) { struct file_view *view = WINE_RB_ENTRY_VALUE( entry, struct file_view, entry ); if (addr < view->base) return -1; if (addr > view->base) return 1; return 0; } /*********************************************************************** * get_prot_str */ static const char *get_prot_str( BYTE prot ) { static char buffer[6]; buffer[0] = (prot & VPROT_COMMITTED) ? 'c' : '-'; buffer[1] = (prot & VPROT_GUARD) ? 'g' : ((prot & VPROT_WRITEWATCH) ? 'H' : '-'); buffer[2] = (prot & VPROT_READ) ? 'r' : '-'; buffer[3] = (prot & VPROT_WRITECOPY) ? 'W' : ((prot & VPROT_WRITE) ? 'w' : '-'); buffer[4] = (prot & VPROT_EXEC) ? 'x' : '-'; buffer[5] = 0; return buffer; } /*********************************************************************** * get_unix_prot * * Convert page protections to protection for mmap/mprotect. */ static int get_unix_prot( BYTE vprot ) { int prot = 0; if ((vprot & VPROT_COMMITTED) && !(vprot & VPROT_GUARD)) { if (vprot & VPROT_READ) prot |= PROT_READ; if (vprot & VPROT_WRITE) prot |= PROT_WRITE | PROT_READ; if (vprot & VPROT_WRITECOPY) prot |= PROT_WRITE | PROT_READ; if (vprot & VPROT_EXEC) prot |= PROT_EXEC | PROT_READ; if (vprot & VPROT_WRITEWATCH) prot &= ~PROT_WRITE; } if (!prot) prot = PROT_NONE; return prot; } /*********************************************************************** * dump_view */ static void dump_view( struct file_view *view ) { UINT i, count; char *addr = view->base; BYTE prot = get_page_vprot( addr ); TRACE( "View: %p - %p", addr, addr + view->size - 1 ); if (view->protect & VPROT_SYSTEM) TRACE( " (builtin image)\n" ); else if (view->protect & SEC_IMAGE) TRACE( " (image)\n" ); else if (view->protect & SEC_FILE) TRACE( " (file)\n" ); else if (view->protect & (SEC_RESERVE | SEC_COMMIT)) TRACE( " (anonymous)\n" ); else TRACE( " (valloc)\n"); for (count = i = 1; i < view->size >> page_shift; i++, count++) { BYTE next = get_page_vprot( addr + (count << page_shift) ); if (next == prot) continue; TRACE( " %p - %p %s\n", addr, addr + (count << page_shift) - 1, get_prot_str(prot) ); addr += (count << page_shift); prot = next; count = 0; } if (count) TRACE( " %p - %p %s\n", addr, addr + (count << page_shift) - 1, get_prot_str(prot) ); } /*********************************************************************** * VIRTUAL_Dump */ #ifdef WINE_VM_DEBUG static void VIRTUAL_Dump(void) { sigset_t sigset; struct file_view *view; TRACE( "Dump of all virtual memory views:\n" ); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); WINE_RB_FOR_EACH_ENTRY( view, &views_tree, struct file_view, entry ) { dump_view( view ); } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } #endif /*********************************************************************** * find_view * * Find the view containing a given address. virtual_mutex must be held by caller. * * PARAMS * addr [I] Address * * RETURNS * View: Success * NULL: Failure */ static struct file_view *find_view( const void *addr, size_t size ) { struct wine_rb_entry *ptr = views_tree.root; if ((const char *)addr + size < (const char *)addr) return NULL; /* overflow */ while (ptr) { struct file_view *view = WINE_RB_ENTRY_VALUE( ptr, struct file_view, entry ); if (view->base > addr) ptr = ptr->left; else if ((const char *)view->base + view->size <= (const char *)addr) ptr = ptr->right; else if ((const char *)view->base + view->size < (const char *)addr + size) break; /* size too large */ else return view; } return NULL; } /*********************************************************************** * zero_bits_win_to_64 * * Convert from Windows hybrid 32bit-based / bitmask to 64bit-based format */ static inline unsigned short zero_bits_win_to_64( ULONG_PTR zero_bits ) { unsigned short zero_bits_64; if (zero_bits == 0) return 0; if (zero_bits < 32) return 32 + zero_bits; zero_bits_64 = 63; #ifdef _WIN64 if (zero_bits >> 32) { zero_bits_64 -= 32; zero_bits >>= 32; } #endif if (zero_bits >> 16) { zero_bits_64 -= 16; zero_bits >>= 16; } if (zero_bits >> 8) { zero_bits_64 -= 8; zero_bits >>= 8; } if (zero_bits >> 4) { zero_bits_64 -= 4; zero_bits >>= 4; } if (zero_bits >> 2) { zero_bits_64 -= 2; zero_bits >>= 2; } if (zero_bits >> 1) { zero_bits_64 -= 1; } return zero_bits_64; } /*********************************************************************** * get_zero_bits_64_mask */ static inline UINT_PTR get_zero_bits_64_mask( USHORT zero_bits_64 ) { return (UINT_PTR)((~(UINT64)0) >> zero_bits_64); } /*********************************************************************** * is_write_watch_range */ static inline BOOL is_write_watch_range( const void *addr, size_t size ) { struct file_view *view = find_view( addr, size ); return view && (view->protect & VPROT_WRITEWATCH); } /*********************************************************************** * find_view_range * * Find the first view overlapping at least part of the specified range. * virtual_mutex must be held by caller. */ static struct file_view *find_view_range( const void *addr, size_t size ) { struct wine_rb_entry *ptr = views_tree.root; while (ptr) { struct file_view *view = WINE_RB_ENTRY_VALUE( ptr, struct file_view, entry ); if ((const char *)view->base >= (const char *)addr + size) ptr = ptr->left; else if ((const char *)view->base + view->size <= (const char *)addr) ptr = ptr->right; else return view; } return NULL; } /*********************************************************************** * find_view_inside_range * * Find first (resp. last, if top_down) view inside a range. * virtual_mutex must be held by caller. */ static struct wine_rb_entry *find_view_inside_range( void **base_ptr, void **end_ptr, int top_down ) { struct wine_rb_entry *first = NULL, *ptr = views_tree.root; void *base = *base_ptr, *end = *end_ptr; /* find the first (resp. last) view inside the range */ while (ptr) { struct file_view *view = WINE_RB_ENTRY_VALUE( ptr, struct file_view, entry ); if ((char *)view->base + view->size >= (char *)end) { end = min( end, view->base ); ptr = ptr->left; } else if (view->base <= base) { base = max( (char *)base, (char *)view->base + view->size ); ptr = ptr->right; } else { first = ptr; ptr = top_down ? ptr->right : ptr->left; } } *base_ptr = base; *end_ptr = end; return first; } /*********************************************************************** * try_map_free_area * * Try mmaping some expected free memory region, eventually stepping and * retrying inside it, and return where it actually succeeded, or NULL. */ static void* try_map_free_area( void *base, void *end, ptrdiff_t step, void *start, size_t size, int unix_prot ) { void *ptr; while (start && base <= start && (char*)start + size <= (char*)end) { if ((ptr = wine_anon_mmap( start, size, unix_prot, 0 )) == start) return start; TRACE( "Found free area is already mapped, start %p.\n", start ); if (ptr != (void *)-1) munmap( ptr, size ); if ((step > 0 && (char *)end - (char *)start < step) || (step < 0 && (char *)start - (char *)base < -step) || step == 0) break; start = (char *)start + step; } return NULL; } /*********************************************************************** * map_free_area * * Find a free area between views inside the specified range and map it. * virtual_mutex must be held by caller. */ static void *map_free_area( void *base, void *end, size_t size, int top_down, int unix_prot ) { struct wine_rb_entry *first = find_view_inside_range( &base, &end, top_down ); ptrdiff_t step = top_down ? -(granularity_mask + 1) : (granularity_mask + 1); void *start; if (top_down) { start = ROUND_ADDR( (char *)end - size, granularity_mask ); if (start >= end || start < base) return NULL; while (first) { struct file_view *view = WINE_RB_ENTRY_VALUE( first, struct file_view, entry ); if ((start = try_map_free_area( (char *)view->base + view->size, (char *)start + size, step, start, size, unix_prot ))) break; start = ROUND_ADDR( (char *)view->base - size, granularity_mask ); /* stop if remaining space is not large enough */ if (!start || start >= end || start < base) return NULL; first = wine_rb_prev( first ); } } else { start = ROUND_ADDR( (char *)base + granularity_mask, granularity_mask ); if (!start || start >= end || (char *)end - (char *)start < size) return NULL; while (first) { struct file_view *view = WINE_RB_ENTRY_VALUE( first, struct file_view, entry ); if ((start = try_map_free_area( start, view->base, step, start, size, unix_prot ))) break; start = ROUND_ADDR( (char *)view->base + view->size + granularity_mask, granularity_mask ); /* stop if remaining space is not large enough */ if (!start || start >= end || (char *)end - (char *)start < size) return NULL; first = wine_rb_next( first ); } } if (!first) return try_map_free_area( base, end, step, start, size, unix_prot ); return start; } /*********************************************************************** * find_reserved_free_area * * Find a free area between views inside the specified range. * virtual_mutex must be held by caller. * The range must be inside the preloader reserved range. */ static void *find_reserved_free_area( void *base, void *end, size_t size, int top_down ) { struct range_entry *range; void *start; base = ROUND_ADDR( (char *)base + granularity_mask, granularity_mask ); end = (char *)ROUND_ADDR( (char *)end - size, granularity_mask ) + size; if (top_down) { start = (char *)end - size; range = free_ranges_lower_bound( start ); assert(range != free_ranges_end && range->end >= start); if ((char *)range->end - (char *)start < size) start = ROUND_ADDR( (char *)range->end - size, granularity_mask ); do { if (start >= end || start < base || (char *)end - (char *)start < size) return NULL; if (start < range->end && start >= range->base && (char *)range->end - (char *)start >= size) break; if (--range < free_ranges) return NULL; start = ROUND_ADDR( (char *)range->end - size, granularity_mask ); } while (1); } else { start = base; range = free_ranges_lower_bound( start ); assert(range != free_ranges_end && range->end >= start); if (start < range->base) start = ROUND_ADDR( (char *)range->base + granularity_mask, granularity_mask ); do { if (start >= end || start < base || (char *)end - (char *)start < size) return NULL; if (start < range->end && start >= range->base && (char *)range->end - (char *)start >= size) break; if (++range == free_ranges_end) return NULL; start = ROUND_ADDR( (char *)range->base + granularity_mask, granularity_mask ); } while (1); } return start; } /*********************************************************************** * add_reserved_area * * Add a reserved area to the list maintained by libwine. * virtual_mutex must be held by caller. */ static void add_reserved_area( void *addr, size_t size ) { TRACE( "adding %p-%p\n", addr, (char *)addr + size ); if (addr < user_space_limit) { /* unmap the part of the area that is below the limit */ assert( (char *)addr + size > (char *)user_space_limit ); munmap( addr, (char *)user_space_limit - (char *)addr ); size -= (char *)user_space_limit - (char *)addr; addr = user_space_limit; } /* blow away existing mappings */ wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED ); mmap_add_reserved_area( addr, size ); } /*********************************************************************** * remove_reserved_area * * Remove a reserved area from the list maintained by libwine. * virtual_mutex must be held by caller. */ static void remove_reserved_area( void *addr, size_t size ) { struct file_view *view; TRACE( "removing %p-%p\n", addr, (char *)addr + size ); mmap_remove_reserved_area( addr, size ); /* unmap areas not covered by an existing view */ WINE_RB_FOR_EACH_ENTRY( view, &views_tree, struct file_view, entry ) { if ((char *)view->base >= (char *)addr + size) break; if ((char *)view->base + view->size <= (char *)addr) continue; if (view->base > addr) munmap( addr, (char *)view->base - (char *)addr ); if ((char *)view->base + view->size > (char *)addr + size) return; size = (char *)addr + size - ((char *)view->base + view->size); addr = (char *)view->base + view->size; } munmap( addr, size ); } struct area_boundary { void *base; size_t size; void *boundary; }; /*********************************************************************** * get_area_boundary_callback * * Get lowest boundary address between reserved area and non-reserved area * in the specified region. If no boundaries are found, result is NULL. * virtual_mutex must be held by caller. */ static int CDECL get_area_boundary_callback( void *start, SIZE_T size, void *arg ) { struct area_boundary *area = arg; void *end = (char *)start + size; area->boundary = NULL; if (area->base >= end) return 0; if ((char *)start >= (char *)area->base + area->size) return 1; if (area->base >= start) { if ((char *)area->base + area->size > (char *)end) { area->boundary = end; return 1; } return 0; } area->boundary = start; return 1; } /*********************************************************************** * is_beyond_limit * * Check if an address range goes beyond a given limit. */ static inline BOOL is_beyond_limit( const void *addr, size_t size, const void *limit ) { return (addr >= limit || (const char *)addr + size > (const char *)limit); } /*********************************************************************** * unmap_area * * Unmap an area, or simply replace it by an empty mapping if it is * in a reserved area. virtual_mutex must be held by caller. */ static inline void unmap_area( void *addr, size_t size ) { switch (mmap_is_in_reserved_area( addr, size )) { case -1: /* partially in a reserved area */ { struct area_boundary area; size_t lower_size; area.base = addr; area.size = size; mmap_enum_reserved_areas( get_area_boundary_callback, &area, 0 ); assert( area.boundary ); lower_size = (char *)area.boundary - (char *)addr; unmap_area( addr, lower_size ); unmap_area( area.boundary, size - lower_size ); break; } case 1: /* in a reserved area */ wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED ); break; default: case 0: /* not in a reserved area */ if (is_beyond_limit( addr, size, user_space_limit )) add_reserved_area( addr, size ); else munmap( addr, size ); break; } } /*********************************************************************** * alloc_view * * Allocate a new view. virtual_mutex must be held by caller. */ static struct file_view *alloc_view(void) { if (next_free_view) { struct file_view *ret = next_free_view; next_free_view = *(struct file_view **)ret; return ret; } if (view_block_start == view_block_end) { void *ptr = wine_anon_mmap( NULL, view_block_size, PROT_READ | PROT_WRITE, 0 ); if (ptr == (void *)-1) return NULL; view_block_start = ptr; view_block_end = view_block_start + view_block_size / sizeof(*view_block_start); } return view_block_start++; } /*********************************************************************** * delete_view * * Deletes a view. virtual_mutex must be held by caller. */ static void delete_view( struct file_view *view ) /* [in] View */ { if (!(view->protect & VPROT_SYSTEM)) unmap_area( view->base, view->size ); set_page_vprot( view->base, view->size, 0 ); if (mmap_is_in_reserved_area( view->base, view->size )) free_ranges_remove_view( view ); wine_rb_remove( &views_tree, &view->entry ); *(struct file_view **)view = next_free_view; next_free_view = view; } /*********************************************************************** * create_view * * Create a view. virtual_mutex must be held by caller. */ static NTSTATUS create_view( struct file_view **view_ret, void *base, size_t size, unsigned int vprot ) { struct file_view *view; int unix_prot = get_unix_prot( vprot ); assert( !((UINT_PTR)base & page_mask) ); assert( !(size & page_mask) ); /* Check for overlapping views. This can happen if the previous view * was a system view that got unmapped behind our back. In that case * we recover by simply deleting it. */ while ((view = find_view_range( base, size ))) { TRACE( "overlapping view %p-%p for %p-%p\n", view->base, (char *)view->base + view->size, base, (char *)base + size ); assert( view->protect & VPROT_SYSTEM ); delete_view( view ); } if (!alloc_pages_vprot( base, size )) return STATUS_NO_MEMORY; /* Create the view structure */ if (!(view = alloc_view())) { FIXME( "out of memory for %p-%p\n", base, (char *)base + size ); return STATUS_NO_MEMORY; } view->base = base; view->size = size; view->protect = vprot; set_page_vprot( base, size, vprot ); wine_rb_put( &views_tree, view->base, &view->entry ); if (mmap_is_in_reserved_area( view->base, view->size )) free_ranges_insert_view( view ); *view_ret = view; if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 ); mprotect( base, size, unix_prot | PROT_EXEC ); } return STATUS_SUCCESS; } /*********************************************************************** * get_win32_prot * * Convert page protections to Win32 flags. */ static DWORD get_win32_prot( BYTE vprot, unsigned int map_prot ) { DWORD ret = VIRTUAL_Win32Flags[vprot & 0x0f]; if (vprot & VPROT_GUARD) ret |= PAGE_GUARD; if (map_prot & SEC_NOCACHE) ret |= PAGE_NOCACHE; return ret; } /*********************************************************************** * get_vprot_flags * * Build page protections from Win32 flags. */ static NTSTATUS get_vprot_flags( DWORD protect, unsigned int *vprot, BOOL image ) { switch(protect & 0xff) { case PAGE_READONLY: *vprot = VPROT_READ; break; case PAGE_READWRITE: if (image) *vprot = VPROT_READ | VPROT_WRITECOPY; else *vprot = VPROT_READ | VPROT_WRITE; break; case PAGE_WRITECOPY: *vprot = VPROT_READ | VPROT_WRITECOPY; break; case PAGE_EXECUTE: *vprot = VPROT_EXEC; break; case PAGE_EXECUTE_READ: *vprot = VPROT_EXEC | VPROT_READ; break; case PAGE_EXECUTE_READWRITE: if (image) *vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITECOPY; else *vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITE; break; case PAGE_EXECUTE_WRITECOPY: *vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITECOPY; break; case PAGE_NOACCESS: *vprot = 0; break; default: return STATUS_INVALID_PAGE_PROTECTION; } if (protect & PAGE_GUARD) *vprot |= VPROT_GUARD; return STATUS_SUCCESS; } /*********************************************************************** * mprotect_exec * * Wrapper for mprotect, adds PROT_EXEC if forced by force_exec_prot */ static inline int mprotect_exec( void *base, size_t size, int unix_prot ) { if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 ); if (!mprotect( base, size, unix_prot | PROT_EXEC )) return 0; /* exec + write may legitimately fail, in that case fall back to write only */ if (!(unix_prot & PROT_WRITE)) return -1; } return mprotect( base, size, unix_prot ); } /*********************************************************************** * mprotect_range * * Call mprotect on a page range, applying the protections from the per-page byte. */ static void mprotect_range( void *base, size_t size, BYTE set, BYTE clear ) { size_t i, count; char *addr = ROUND_ADDR( base, page_mask ); int prot, next; size = ROUND_SIZE( base, size ); prot = get_unix_prot( (get_page_vprot( addr ) & ~clear ) | set ); for (count = i = 1; i < size >> page_shift; i++, count++) { next = get_unix_prot( (get_page_vprot( addr + (count << page_shift) ) & ~clear) | set ); if (next == prot) continue; mprotect_exec( addr, count << page_shift, prot ); addr += count << page_shift; prot = next; count = 0; } if (count) mprotect_exec( addr, count << page_shift, prot ); } /*********************************************************************** * set_vprot * * Change the protection of a range of pages. */ static BOOL set_vprot( struct file_view *view, void *base, size_t size, BYTE vprot ) { int unix_prot = get_unix_prot(vprot); if (view->protect & VPROT_WRITEWATCH) { /* each page may need different protections depending on write watch flag */ set_page_vprot_bits( base, size, vprot & ~VPROT_WRITEWATCH, ~vprot & ~VPROT_WRITEWATCH ); mprotect_range( base, size, 0, 0 ); return TRUE; } /* if setting stack guard pages, store the permissions first, as the guard may be * triggered at any point after mprotect and change the permissions again */ if ((vprot & VPROT_GUARD) && (base >= NtCurrentTeb()->DeallocationStack) && (base < NtCurrentTeb()->Tib.StackBase)) { set_page_vprot( base, size, vprot ); mprotect( base, size, unix_prot ); return TRUE; } if (mprotect_exec( base, size, unix_prot )) /* FIXME: last error */ return FALSE; set_page_vprot( base, size, vprot ); return TRUE; } /*********************************************************************** * set_protection * * Set page protections on a range of pages */ static NTSTATUS set_protection( struct file_view *view, void *base, SIZE_T size, ULONG protect ) { unsigned int vprot; NTSTATUS status; if ((status = get_vprot_flags( protect, &vprot, view->protect & SEC_IMAGE ))) return status; if (is_view_valloc( view )) { if (vprot & VPROT_WRITECOPY) return STATUS_INVALID_PAGE_PROTECTION; } else { BYTE access = vprot & (VPROT_READ | VPROT_WRITE | VPROT_EXEC); if ((view->protect & access) != access) return STATUS_INVALID_PAGE_PROTECTION; } if (!set_vprot( view, base, size, vprot | VPROT_COMMITTED )) return STATUS_ACCESS_DENIED; return STATUS_SUCCESS; } /*********************************************************************** * update_write_watches */ static void update_write_watches( void *base, size_t size, size_t accessed_size ) { TRACE( "updating watch %p-%p-%p\n", base, (char *)base + accessed_size, (char *)base + size ); /* clear write watch flag on accessed pages */ set_page_vprot_bits( base, accessed_size, 0, VPROT_WRITEWATCH ); /* restore page protections on the entire range */ mprotect_range( base, size, 0, 0 ); } /*********************************************************************** * reset_write_watches * * Reset write watches in a memory range. */ static void reset_write_watches( void *base, SIZE_T size ) { set_page_vprot_bits( base, size, VPROT_WRITEWATCH, 0 ); mprotect_range( base, size, 0, 0 ); } /*********************************************************************** * unmap_extra_space * * Release the extra memory while keeping the range starting on the granularity boundary. */ static inline void *unmap_extra_space( void *ptr, size_t total_size, size_t wanted_size ) { if ((ULONG_PTR)ptr & granularity_mask) { size_t extra = granularity_mask + 1 - ((ULONG_PTR)ptr & granularity_mask); munmap( ptr, extra ); ptr = (char *)ptr + extra; total_size -= extra; } if (total_size > wanted_size) munmap( (char *)ptr + wanted_size, total_size - wanted_size ); return ptr; } struct alloc_area { size_t size; int top_down; void *limit; void *result; }; /*********************************************************************** * alloc_reserved_area_callback * * Try to map some space inside a reserved area. Callback for mmap_enum_reserved_areas. */ static int CDECL alloc_reserved_area_callback( void *start, SIZE_T size, void *arg ) { struct alloc_area *alloc = arg; void *end = (char *)start + size; if (start < address_space_start) start = address_space_start; if (is_beyond_limit( start, size, alloc->limit )) end = alloc->limit; if (start >= end) return 0; /* make sure we don't touch the preloader reserved range */ if (preload_reserve_end >= start) { if (preload_reserve_end >= end) { if (preload_reserve_start <= start) return 0; /* no space in that area */ if (preload_reserve_start < end) end = preload_reserve_start; } else if (preload_reserve_start <= start) start = preload_reserve_end; else { /* range is split in two by the preloader reservation, try first part */ if ((alloc->result = find_reserved_free_area( start, preload_reserve_start, alloc->size, alloc->top_down ))) return 1; /* then fall through to try second part */ start = preload_reserve_end; } } if ((alloc->result = find_reserved_free_area( start, end, alloc->size, alloc->top_down ))) return 1; return 0; } /*********************************************************************** * map_fixed_area * * mmap the fixed memory area. * virtual_mutex must be held by caller. */ static NTSTATUS map_fixed_area( void *base, size_t size, unsigned int vprot ) { void *ptr; switch (mmap_is_in_reserved_area( base, size )) { case -1: /* partially in a reserved area */ { NTSTATUS status; struct area_boundary area; size_t lower_size; area.base = base; area.size = size; mmap_enum_reserved_areas( get_area_boundary_callback, &area, 0 ); assert( area.boundary ); lower_size = (char *)area.boundary - (char *)base; status = map_fixed_area( base, lower_size, vprot ); if (status == STATUS_SUCCESS) { status = map_fixed_area( area.boundary, size - lower_size, vprot); if (status != STATUS_SUCCESS) unmap_area( base, lower_size ); } return status; } case 0: /* not in a reserved area, do a normal allocation */ if ((ptr = wine_anon_mmap( base, size, get_unix_prot(vprot), 0 )) == (void *)-1) { if (errno == ENOMEM) return STATUS_NO_MEMORY; return STATUS_INVALID_PARAMETER; } if (ptr != base) { /* We couldn't get the address we wanted */ if (is_beyond_limit( ptr, size, user_space_limit )) add_reserved_area( ptr, size ); else munmap( ptr, size ); return STATUS_CONFLICTING_ADDRESSES; } break; default: case 1: /* in a reserved area, make sure the address is available */ if (find_view_range( base, size )) return STATUS_CONFLICTING_ADDRESSES; /* replace the reserved area by our mapping */ if ((ptr = wine_anon_mmap( base, size, get_unix_prot(vprot), MAP_FIXED )) != base) return STATUS_INVALID_PARAMETER; break; } if (is_beyond_limit( ptr, size, working_set_limit )) working_set_limit = address_space_limit; return STATUS_SUCCESS; } /*********************************************************************** * map_view * * Create a view and mmap the corresponding memory area. * virtual_mutex must be held by caller. */ static NTSTATUS map_view( struct file_view **view_ret, void *base, size_t size, int top_down, unsigned int vprot, unsigned short zero_bits_64 ) { void *ptr; NTSTATUS status; if (base) { if (is_beyond_limit( base, size, address_space_limit )) return STATUS_WORKING_SET_LIMIT_RANGE; status = map_fixed_area( base, size, vprot ); if (status != STATUS_SUCCESS) return status; ptr = base; } else { size_t view_size = size + granularity_mask + 1; struct alloc_area alloc; alloc.size = size; alloc.top_down = top_down; alloc.limit = (void*)(get_zero_bits_64_mask( zero_bits_64 ) & (UINT_PTR)user_space_limit); if (mmap_enum_reserved_areas( alloc_reserved_area_callback, &alloc, top_down )) { ptr = alloc.result; TRACE( "got mem in reserved area %p-%p\n", ptr, (char *)ptr + size ); if (wine_anon_mmap( ptr, size, get_unix_prot(vprot), MAP_FIXED ) != ptr) return STATUS_INVALID_PARAMETER; goto done; } if (zero_bits_64) { if (!(ptr = map_free_area( address_space_start, alloc.limit, size, top_down, get_unix_prot(vprot) ))) return STATUS_NO_MEMORY; TRACE( "got mem with map_free_area %p-%p\n", ptr, (char *)ptr + size ); goto done; } for (;;) { if ((ptr = wine_anon_mmap( NULL, view_size, get_unix_prot(vprot), 0 )) == (void *)-1) { if (errno == ENOMEM) return STATUS_NO_MEMORY; return STATUS_INVALID_PARAMETER; } TRACE( "got mem with anon mmap %p-%p\n", ptr, (char *)ptr + size ); /* if we got something beyond the user limit, unmap it and retry */ if (is_beyond_limit( ptr, view_size, user_space_limit )) add_reserved_area( ptr, view_size ); else break; } ptr = unmap_extra_space( ptr, view_size, size ); } done: status = create_view( view_ret, ptr, size, vprot ); if (status != STATUS_SUCCESS) unmap_area( ptr, size ); return status; } /*********************************************************************** * map_file_into_view * * Wrapper for mmap() to map a file into a view, falling back to read if mmap fails. * virtual_mutex must be held by caller. */ static NTSTATUS map_file_into_view( struct file_view *view, int fd, size_t start, size_t size, off_t offset, unsigned int vprot, BOOL removable ) { void *ptr; int prot = get_unix_prot( vprot | VPROT_COMMITTED /* make sure it is accessible */ ); unsigned int flags = MAP_FIXED | ((vprot & VPROT_WRITECOPY) ? MAP_PRIVATE : MAP_SHARED); assert( start < view->size ); assert( start + size <= view->size ); if (force_exec_prot && (vprot & VPROT_READ)) { TRACE( "forcing exec permission on mapping %p-%p\n", (char *)view->base + start, (char *)view->base + start + size - 1 ); prot |= PROT_EXEC; } /* only try mmap if media is not removable (or if we require write access) */ if (!removable || (flags & MAP_SHARED)) { if (mmap( (char *)view->base + start, size, prot, flags, fd, offset ) != (void *)-1) goto done; switch (errno) { case EINVAL: /* file offset is not page-aligned, fall back to read() */ if (flags & MAP_SHARED) return STATUS_INVALID_PARAMETER; break; case ENOEXEC: case ENODEV: /* filesystem doesn't support mmap(), fall back to read() */ if (vprot & VPROT_WRITE) { ERR( "shared writable mmap not supported, broken filesystem?\n" ); return STATUS_NOT_SUPPORTED; } break; case EACCES: case EPERM: /* noexec filesystem, fall back to read() */ if (flags & MAP_SHARED) { if (prot & PROT_EXEC) ERR( "failed to set PROT_EXEC on file map, noexec filesystem?\n" ); return STATUS_ACCESS_DENIED; } if (prot & PROT_EXEC) WARN( "failed to set PROT_EXEC on file map, noexec filesystem?\n" ); break; default: return STATUS_NO_MEMORY; } } /* Reserve the memory with an anonymous mmap */ ptr = wine_anon_mmap( (char *)view->base + start, size, PROT_READ | PROT_WRITE, MAP_FIXED ); if (ptr == (void *)-1) return STATUS_NO_MEMORY; /* Now read in the file */ pread( fd, ptr, size, offset ); if (prot != (PROT_READ|PROT_WRITE)) mprotect( ptr, size, prot ); /* Set the right protection */ done: set_page_vprot( (char *)view->base + start, size, vprot ); return STATUS_SUCCESS; } /*********************************************************************** * get_committed_size * * Get the size of the committed range starting at base. * Also return the protections for the first page. */ static SIZE_T get_committed_size( struct file_view *view, void *base, BYTE *vprot ) { SIZE_T i, start; start = ((char *)base - (char *)view->base) >> page_shift; *vprot = get_page_vprot( base ); if (view->protect & SEC_RESERVE) { SIZE_T ret = 0; SERVER_START_REQ( get_mapping_committed_range ) { req->base = wine_server_client_ptr( view->base ); req->offset = start << page_shift; if (!wine_server_call( req )) { ret = reply->size; if (reply->committed) { *vprot |= VPROT_COMMITTED; set_page_vprot_bits( base, ret, VPROT_COMMITTED, 0 ); } } } SERVER_END_REQ; return ret; } for (i = start + 1; i < view->size >> page_shift; i++) if ((*vprot ^ get_page_vprot( (char *)view->base + (i << page_shift) )) & VPROT_COMMITTED) break; return (i - start) << page_shift; } /*********************************************************************** * decommit_view * * Decommit some pages of a given view. * virtual_mutex must be held by caller. */ static NTSTATUS decommit_pages( struct file_view *view, size_t start, size_t size ) { if (wine_anon_mmap( (char *)view->base + start, size, PROT_NONE, MAP_FIXED ) != (void *)-1) { set_page_vprot_bits( (char *)view->base + start, size, 0, VPROT_COMMITTED ); return STATUS_SUCCESS; } return STATUS_NO_MEMORY; } /*********************************************************************** * allocate_dos_memory * * Allocate the DOS memory range. */ static NTSTATUS allocate_dos_memory( struct file_view **view, unsigned int vprot ) { size_t size; void *addr = NULL; void * const low_64k = (void *)0x10000; const size_t dosmem_size = 0x110000; int unix_prot = get_unix_prot( vprot ); /* check for existing view */ if (find_view_range( 0, dosmem_size )) return STATUS_CONFLICTING_ADDRESSES; /* check without the first 64K */ if (mmap_is_in_reserved_area( low_64k, dosmem_size - 0x10000 ) != 1) { addr = wine_anon_mmap( low_64k, dosmem_size - 0x10000, unix_prot, 0 ); if (addr != low_64k) { if (addr != (void *)-1) munmap( addr, dosmem_size - 0x10000 ); return map_view( view, NULL, dosmem_size, FALSE, vprot, 0 ); } } /* now try to allocate the low 64K too */ if (mmap_is_in_reserved_area( NULL, 0x10000 ) != 1) { addr = wine_anon_mmap( (void *)page_size, 0x10000 - page_size, unix_prot, 0 ); if (addr == (void *)page_size) { if (!wine_anon_mmap( NULL, page_size, unix_prot, MAP_FIXED )) { addr = NULL; TRACE( "successfully mapped low 64K range\n" ); } else TRACE( "failed to map page 0\n" ); } else { if (addr != (void *)-1) munmap( addr, 0x10000 - page_size ); addr = low_64k; TRACE( "failed to map low 64K range\n" ); } } /* now reserve the whole range */ size = (char *)dosmem_size - (char *)addr; wine_anon_mmap( addr, size, unix_prot, MAP_FIXED ); return create_view( view, addr, size, vprot ); } /*********************************************************************** * map_pe_header * * Map the header of a PE file into memory. */ static NTSTATUS map_pe_header( void *ptr, size_t size, int fd, BOOL *removable ) { if (!size) return STATUS_INVALID_IMAGE_FORMAT; if (!*removable) { if (mmap( ptr, size, PROT_READ|PROT_WRITE|PROT_EXEC, MAP_FIXED|MAP_PRIVATE, fd, 0 ) != (void *)-1) return STATUS_SUCCESS; switch (errno) { case EPERM: case EACCES: WARN( "noexec file system, falling back to read\n" ); break; case ENOEXEC: case ENODEV: WARN( "file system doesn't support mmap, falling back to read\n" ); break; default: return STATUS_NO_MEMORY; } *removable = TRUE; } pread( fd, ptr, size, 0 ); return STATUS_SUCCESS; /* page protections will be updated later */ } /*********************************************************************** * map_image_into_view * * Map an executable (PE format) image into an existing view. * virtual_mutex must be held by caller. */ static NTSTATUS map_image_into_view( struct file_view *view, int fd, void *orig_base, SIZE_T header_size, ULONG image_flags, int shared_fd, BOOL removable ) { IMAGE_DOS_HEADER *dos; IMAGE_NT_HEADERS *nt; IMAGE_SECTION_HEADER sections[96]; IMAGE_SECTION_HEADER *sec; IMAGE_DATA_DIRECTORY *imports; NTSTATUS status = STATUS_CONFLICTING_ADDRESSES; int i; off_t pos; struct stat st; char *header_end, *header_start; char *ptr = view->base; SIZE_T total_size = view->size; TRACE_(module)( "mapped PE file at %p-%p\n", ptr, ptr + total_size ); /* map the header */ fstat( fd, &st ); header_size = min( header_size, st.st_size ); if ((status = map_pe_header( view->base, header_size, fd, &removable ))) return status; status = STATUS_INVALID_IMAGE_FORMAT; /* generic error */ dos = (IMAGE_DOS_HEADER *)ptr; nt = (IMAGE_NT_HEADERS *)(ptr + dos->e_lfanew); header_end = ptr + ROUND_SIZE( 0, header_size ); memset( ptr + header_size, 0, header_end - (ptr + header_size) ); if ((char *)(nt + 1) > header_end) return status; header_start = (char*)&nt->OptionalHeader+nt->FileHeader.SizeOfOptionalHeader; if (nt->FileHeader.NumberOfSections > ARRAY_SIZE( sections )) return status; if (header_start + sizeof(*sections) * nt->FileHeader.NumberOfSections > header_end) return status; /* Some applications (e.g. the Steam version of Borderlands) map over the top of the section headers, * copying the headers into local memory is necessary to properly load such applications. */ memcpy(sections, header_start, sizeof(*sections) * nt->FileHeader.NumberOfSections); sec = sections; imports = nt->OptionalHeader.DataDirectory + IMAGE_DIRECTORY_ENTRY_IMPORT; if (!imports->Size || !imports->VirtualAddress) imports = NULL; /* check for non page-aligned binary */ if (image_flags & IMAGE_FLAGS_ImageMappedFlat) { /* unaligned sections, this happens for native subsystem binaries */ /* in that case Windows simply maps in the whole file */ total_size = min( total_size, ROUND_SIZE( 0, st.st_size )); if (map_file_into_view( view, fd, 0, total_size, 0, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, removable ) != STATUS_SUCCESS) return status; /* check that all sections are loaded at the right offset */ if (nt->OptionalHeader.FileAlignment != nt->OptionalHeader.SectionAlignment) return status; for (i = 0; i < nt->FileHeader.NumberOfSections; i++) { if (sec[i].VirtualAddress != sec[i].PointerToRawData) return status; /* Windows refuses to load in that case too */ } /* set the image protections */ set_vprot( view, ptr, total_size, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY | VPROT_EXEC ); /* no relocations are performed on non page-aligned binaries */ return STATUS_SUCCESS; } /* map all the sections */ for (i = pos = 0; i < nt->FileHeader.NumberOfSections; i++, sec++) { static const SIZE_T sector_align = 0x1ff; SIZE_T map_size, file_start, file_size, end; if (!sec->Misc.VirtualSize) map_size = ROUND_SIZE( 0, sec->SizeOfRawData ); else map_size = ROUND_SIZE( 0, sec->Misc.VirtualSize ); /* file positions are rounded to sector boundaries regardless of OptionalHeader.FileAlignment */ file_start = sec->PointerToRawData & ~sector_align; file_size = (sec->SizeOfRawData + (sec->PointerToRawData & sector_align) + sector_align) & ~sector_align; if (file_size > map_size) file_size = map_size; /* a few sanity checks */ end = sec->VirtualAddress + ROUND_SIZE( sec->VirtualAddress, map_size ); if (sec->VirtualAddress > total_size || end > total_size || end < sec->VirtualAddress) { WARN_(module)( "Section %.8s too large (%x+%lx/%lx)\n", sec->Name, sec->VirtualAddress, map_size, total_size ); return status; } if ((sec->Characteristics & IMAGE_SCN_MEM_SHARED) && (sec->Characteristics & IMAGE_SCN_MEM_WRITE)) { TRACE_(module)( "mapping shared section %.8s at %p off %x (%x) size %lx (%lx) flags %x\n", sec->Name, ptr + sec->VirtualAddress, sec->PointerToRawData, (int)pos, file_size, map_size, sec->Characteristics ); if (map_file_into_view( view, shared_fd, sec->VirtualAddress, map_size, pos, VPROT_COMMITTED | VPROT_READ | VPROT_WRITE, FALSE ) != STATUS_SUCCESS) { ERR_(module)( "Could not map shared section %.8s\n", sec->Name ); return status; } /* check if the import directory falls inside this section */ if (imports && imports->VirtualAddress >= sec->VirtualAddress && imports->VirtualAddress < sec->VirtualAddress + map_size) { UINT_PTR base = imports->VirtualAddress & ~page_mask; UINT_PTR end = base + ROUND_SIZE( imports->VirtualAddress, imports->Size ); if (end > sec->VirtualAddress + map_size) end = sec->VirtualAddress + map_size; if (end > base) map_file_into_view( view, shared_fd, base, end - base, pos + (base - sec->VirtualAddress), VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, FALSE ); } pos += map_size; continue; } TRACE_(module)( "mapping section %.8s at %p off %x size %x virt %x flags %x\n", sec->Name, ptr + sec->VirtualAddress, sec->PointerToRawData, sec->SizeOfRawData, sec->Misc.VirtualSize, sec->Characteristics ); if (!sec->PointerToRawData || !file_size) continue; /* Note: if the section is not aligned properly map_file_into_view will magically * fall back to read(), so we don't need to check anything here. */ end = file_start + file_size; if (sec->PointerToRawData >= st.st_size || end > ((st.st_size + sector_align) & ~sector_align) || end < file_start || map_file_into_view( view, fd, sec->VirtualAddress, file_size, file_start, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, removable ) != STATUS_SUCCESS) { ERR_(module)( "Could not map section %.8s, file probably truncated\n", sec->Name ); return status; } if (file_size & page_mask) { end = ROUND_SIZE( 0, file_size ); if (end > map_size) end = map_size; TRACE_(module)("clearing %p - %p\n", ptr + sec->VirtualAddress + file_size, ptr + sec->VirtualAddress + end ); memset( ptr + sec->VirtualAddress + file_size, 0, end - file_size ); } } /* set the image protections */ set_vprot( view, ptr, ROUND_SIZE( 0, header_size ), VPROT_COMMITTED | VPROT_READ ); sec = sections; for (i = 0; i < nt->FileHeader.NumberOfSections; i++, sec++) { SIZE_T size; BYTE vprot = VPROT_COMMITTED; if (sec->Misc.VirtualSize) size = ROUND_SIZE( sec->VirtualAddress, sec->Misc.VirtualSize ); else size = ROUND_SIZE( sec->VirtualAddress, sec->SizeOfRawData ); if (sec->Characteristics & IMAGE_SCN_MEM_READ) vprot |= VPROT_READ; if (sec->Characteristics & IMAGE_SCN_MEM_WRITE) vprot |= VPROT_WRITECOPY; if (sec->Characteristics & IMAGE_SCN_MEM_EXECUTE) vprot |= VPROT_EXEC; /* Dumb game crack lets the AOEP point into a data section. Adjust. */ if ((nt->OptionalHeader.AddressOfEntryPoint >= sec->VirtualAddress) && (nt->OptionalHeader.AddressOfEntryPoint < sec->VirtualAddress + size)) vprot |= VPROT_EXEC; if (!set_vprot( view, ptr + sec->VirtualAddress, size, vprot ) && (vprot & VPROT_EXEC)) ERR( "failed to set %08x protection on section %.8s, noexec filesystem?\n", sec->Characteristics, sec->Name ); } #ifdef VALGRIND_LOAD_PDB_DEBUGINFO VALGRIND_LOAD_PDB_DEBUGINFO(fd, ptr, total_size, ptr - (char *)orig_base); #endif return STATUS_SUCCESS; } /*********************************************************************** * virtual_map_section * * Map a file section into memory. */ NTSTATUS CDECL virtual_map_section( HANDLE handle, PVOID *addr_ptr, unsigned short zero_bits_64, SIZE_T commit_size, const LARGE_INTEGER *offset_ptr, SIZE_T *size_ptr, ULONG alloc_type, ULONG protect, pe_image_info_t *image_info ) { NTSTATUS res; mem_size_t full_size; ACCESS_MASK access; SIZE_T size; void *base; int unix_handle = -1, needs_close; int shared_fd = -1, shared_needs_close = 0; unsigned int vprot, sec_flags; struct file_view *view; HANDLE shared_file; LARGE_INTEGER offset; sigset_t sigset; offset.QuadPart = offset_ptr ? offset_ptr->QuadPart : 0; switch(protect) { case PAGE_NOACCESS: case PAGE_READONLY: case PAGE_WRITECOPY: access = SECTION_MAP_READ; break; case PAGE_READWRITE: access = SECTION_MAP_WRITE; break; case PAGE_EXECUTE: case PAGE_EXECUTE_READ: case PAGE_EXECUTE_WRITECOPY: access = SECTION_MAP_READ | SECTION_MAP_EXECUTE; break; case PAGE_EXECUTE_READWRITE: access = SECTION_MAP_WRITE | SECTION_MAP_EXECUTE; break; default: return STATUS_INVALID_PAGE_PROTECTION; } SERVER_START_REQ( get_mapping_info ) { req->handle = wine_server_obj_handle( handle ); req->access = access; wine_server_set_reply( req, image_info, sizeof(*image_info) ); res = wine_server_call( req ); sec_flags = reply->flags; full_size = reply->size; shared_file = wine_server_ptr_handle( reply->shared_file ); } SERVER_END_REQ; if (res) return res; if ((res = server_get_unix_fd( handle, 0, &unix_handle, &needs_close, NULL, NULL ))) { if (shared_file) NtClose( shared_file ); return res; } if (shared_file && ((res = server_get_unix_fd( shared_file, FILE_READ_DATA|FILE_WRITE_DATA, &shared_fd, &shared_needs_close, NULL, NULL )))) { NtClose( shared_file ); if (needs_close) close( unix_handle ); return res; } res = STATUS_INVALID_PARAMETER; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (sec_flags & SEC_IMAGE) { base = wine_server_get_ptr( image_info->base ); if ((ULONG_PTR)base != image_info->base) base = NULL; size = image_info->map_size; vprot = SEC_IMAGE | SEC_FILE | VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY; if ((char *)base >= (char *)address_space_start) /* make sure the DOS area remains free */ res = map_view( &view, base, size, alloc_type & MEM_TOP_DOWN, vprot, zero_bits_64 ); if (res) res = map_view( &view, NULL, size, alloc_type & MEM_TOP_DOWN, vprot, zero_bits_64 ); if (res) goto done; res = map_image_into_view( view, unix_handle, base, image_info->header_size, image_info->image_flags, shared_fd, needs_close ); } else { base = *addr_ptr; if (offset.QuadPart >= full_size) goto done; if (*size_ptr) { size = *size_ptr; if (size > full_size - offset.QuadPart) { res = STATUS_INVALID_VIEW_SIZE; goto done; } } else { size = full_size - offset.QuadPart; if (size != full_size - offset.QuadPart) /* truncated */ { WARN( "Files larger than 4Gb (%s) not supported on this platform\n", wine_dbgstr_longlong(full_size) ); goto done; } } if (!(size = ROUND_SIZE( 0, size ))) goto done; /* wrap-around */ get_vprot_flags( protect, &vprot, FALSE ); vprot |= sec_flags; if (!(sec_flags & SEC_RESERVE)) vprot |= VPROT_COMMITTED; res = map_view( &view, base, size, alloc_type & MEM_TOP_DOWN, vprot, zero_bits_64 ); if (res) goto done; TRACE( "handle=%p size=%lx offset=%x%08x\n", handle, size, offset.u.HighPart, offset.u.LowPart ); res = map_file_into_view( view, unix_handle, 0, size, offset.QuadPart, vprot, needs_close ); if (res) ERR( "mapping %p %lx %x%08x failed\n", view->base, size, offset.u.HighPart, offset.u.LowPart ); } if (res == STATUS_SUCCESS) { SERVER_START_REQ( map_view ) { req->mapping = wine_server_obj_handle( handle ); req->access = access; req->base = wine_server_client_ptr( view->base ); req->size = size; req->start = offset.QuadPart; res = wine_server_call( req ); } SERVER_END_REQ; } if (res >= 0) { *addr_ptr = view->base; *size_ptr = size; VIRTUAL_DEBUG_DUMP_VIEW( view ); } else delete_view( view ); done: server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if (needs_close) close( unix_handle ); if (shared_needs_close) close( shared_fd ); if (shared_file) NtClose( shared_file ); return res; } struct alloc_virtual_heap { void *base; size_t size; }; /* callback for mmap_enum_reserved_areas to allocate space for the virtual heap */ static int CDECL alloc_virtual_heap( void *base, SIZE_T size, void *arg ) { struct alloc_virtual_heap *alloc = arg; if (is_beyond_limit( base, size, address_space_limit )) address_space_limit = (char *)base + size; if (size < alloc->size) return 0; if (is_win64 && base < (void *)0x80000000) return 0; alloc->base = wine_anon_mmap( (char *)base + size - alloc->size, alloc->size, PROT_READ|PROT_WRITE, MAP_FIXED ); return (alloc->base != (void *)-1); } /*********************************************************************** * virtual_init */ void virtual_init(void) { const struct preload_info **preload_info = dlsym( RTLD_DEFAULT, "wine_main_preload_info" ); const char *preload = getenv( "WINEPRELOADRESERVE" ); struct alloc_virtual_heap alloc_views; size_t size; int i; pthread_mutexattr_t attr; pthread_mutexattr_init( &attr ); pthread_mutexattr_settype( &attr, PTHREAD_MUTEX_RECURSIVE ); pthread_mutex_init( &virtual_mutex, &attr ); pthread_mutexattr_destroy( &attr ); if (preload_info && *preload_info) for (i = 0; (*preload_info)[i].size; i++) mmap_add_reserved_area( (*preload_info)[i].addr, (*preload_info)[i].size ); mmap_init( preload_info ? *preload_info : NULL ); if ((preload = getenv("WINEPRELOADRESERVE"))) { unsigned long start, end; if (sscanf( preload, "%lx-%lx", &start, &end ) == 2) { preload_reserve_start = (void *)start; preload_reserve_end = (void *)end; /* some apps start inside the DOS area */ if (preload_reserve_start) address_space_start = min( address_space_start, preload_reserve_start ); } } size = ROUND_SIZE( 0, sizeof(TEB) ) + max( MINSIGSTKSZ, 8192 ); /* find the first power of two not smaller than size */ signal_stack_align = page_shift; while ((1u << signal_stack_align) < size) signal_stack_align++; signal_stack_mask = (1 << signal_stack_align) - 1; signal_stack_size = (1 << signal_stack_align) - ROUND_SIZE( 0, sizeof(TEB) ); /* try to find space in a reserved area for the views and pages protection table */ #ifdef _WIN64 pages_vprot_size = ((size_t)address_space_limit >> page_shift >> pages_vprot_shift) + 1; alloc_views.size = 2 * view_block_size + pages_vprot_size * sizeof(*pages_vprot); #else alloc_views.size = 2 * view_block_size + (1U << (32 - page_shift)); #endif if (mmap_enum_reserved_areas( alloc_virtual_heap, &alloc_views, 1 )) mmap_remove_reserved_area( alloc_views.base, alloc_views.size ); else alloc_views.base = wine_anon_mmap( NULL, alloc_views.size, PROT_READ | PROT_WRITE, 0 ); assert( alloc_views.base != (void *)-1 ); view_block_start = alloc_views.base; view_block_end = view_block_start + view_block_size / sizeof(*view_block_start); free_ranges = (void *)((char *)alloc_views.base + view_block_size); pages_vprot = (void *)((char *)alloc_views.base + 2 * view_block_size); wine_rb_init( &views_tree, compare_view ); free_ranges[0].base = (void *)0; free_ranges[0].end = (void *)~0; free_ranges_end = free_ranges + 1; /* make the DOS area accessible (except the low 64K) to hide bugs in broken apps like Excel 2003 */ size = (char *)address_space_start - (char *)0x10000; if (size && mmap_is_in_reserved_area( (void*)0x10000, size ) == 1) wine_anon_mmap( (void *)0x10000, size, PROT_READ | PROT_WRITE, MAP_FIXED ); } /*********************************************************************** * virtual_map_ntdll * * Map ntdll, used instead of virtual_map_section() because some things are not initialized yet. */ NTSTATUS virtual_map_ntdll( int fd, void **module ) { IMAGE_DOS_HEADER dos; IMAGE_NT_HEADERS nt; NTSTATUS status; SIZE_T size; void *base; unsigned int vprot; struct file_view *view; /* load the headers */ size = pread( fd, &dos, sizeof(dos), 0 ); if (size < sizeof(dos)) return STATUS_INVALID_IMAGE_FORMAT; if (dos.e_magic != IMAGE_DOS_SIGNATURE) return STATUS_INVALID_IMAGE_FORMAT; size = pread( fd, &nt, sizeof(nt), dos.e_lfanew ); if (size < sizeof(nt)) return STATUS_INVALID_IMAGE_PROTECT; if (nt.Signature != IMAGE_NT_SIGNATURE) return STATUS_INVALID_IMAGE_FORMAT; if (nt.OptionalHeader.Magic != IMAGE_NT_OPTIONAL_HDR_MAGIC) return STATUS_INVALID_IMAGE_FORMAT; #ifdef __i386__ if (nt.FileHeader.Machine != IMAGE_FILE_MACHINE_I386) return STATUS_INVALID_IMAGE_FORMAT; #elif defined(__x86_64__) if (nt.FileHeader.Machine != IMAGE_FILE_MACHINE_AMD64) return STATUS_INVALID_IMAGE_FORMAT; #elif defined(__arm__) if (nt.FileHeader.Machine != IMAGE_FILE_MACHINE_ARM && nt.FileHeader.Machine != IMAGE_FILE_MACHINE_THUMB && nt.FileHeader.Machine != IMAGE_FILE_MACHINE_ARMNT) return STATUS_INVALID_IMAGE_FORMAT; #elif defined(__aarch64__) if (nt.FileHeader.Machine != IMAGE_FILE_MACHINE_ARM64) return STATUS_INVALID_IMAGE_FORMAT; #endif base = (void *)nt.OptionalHeader.ImageBase; size = ROUND_SIZE( 0, nt.OptionalHeader.SizeOfImage ); vprot = SEC_IMAGE | SEC_FILE | VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY; status = map_view( &view, base, size, FALSE, vprot, 0 ); if (status == STATUS_CONFLICTING_ADDRESSES) ERR( "couldn't load ntdll at preferred address %p\n", base ); if (status) return status; *module = view->base; return map_image_into_view( view, fd, base, nt.OptionalHeader.SizeOfHeaders, 0, -1, FALSE ); } /*********************************************************************** * get_system_affinity_mask */ ULONG_PTR get_system_affinity_mask(void) { ULONG num_cpus = NtCurrentTeb()->Peb->NumberOfProcessors; if (num_cpus >= sizeof(ULONG_PTR) * 8) return ~(ULONG_PTR)0; return ((ULONG_PTR)1 << num_cpus) - 1; } /*********************************************************************** * virtual_get_system_info */ void virtual_get_system_info( SYSTEM_BASIC_INFORMATION *info ) { #ifdef HAVE_SYSINFO struct sysinfo sinfo; #endif info->unknown = 0; info->KeMaximumIncrement = 0; /* FIXME */ info->PageSize = page_size; info->MmLowestPhysicalPage = 1; info->MmHighestPhysicalPage = 0x7fffffff / page_size; #ifdef HAVE_SYSINFO if (!sysinfo(&sinfo)) { ULONG64 total = (ULONG64)sinfo.totalram * sinfo.mem_unit; info->MmHighestPhysicalPage = max(1, total / page_size); } #endif info->MmNumberOfPhysicalPages = info->MmHighestPhysicalPage - info->MmLowestPhysicalPage; info->AllocationGranularity = granularity_mask + 1; info->LowestUserAddress = (void *)0x10000; info->HighestUserAddress = (char *)user_space_limit - 1; info->ActiveProcessorsAffinityMask = get_system_affinity_mask(); info->NumberOfProcessors = NtCurrentTeb()->Peb->NumberOfProcessors; } /*********************************************************************** * virtual_create_builtin_view */ NTSTATUS virtual_create_builtin_view( void *module ) { NTSTATUS status; sigset_t sigset; IMAGE_DOS_HEADER *dos = module; IMAGE_NT_HEADERS *nt = (IMAGE_NT_HEADERS *)((char *)dos + dos->e_lfanew); SIZE_T size = nt->OptionalHeader.SizeOfImage; IMAGE_SECTION_HEADER *sec; struct file_view *view; void *base; int i; size = ROUND_SIZE( module, size ); base = ROUND_ADDR( module, page_mask ); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); status = create_view( &view, base, size, SEC_IMAGE | SEC_FILE | VPROT_SYSTEM | VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY | VPROT_EXEC ); if (!status) { TRACE( "created %p-%p\n", base, (char *)base + size ); /* The PE header is always read-only, no write, no execute. */ set_page_vprot( base, page_size, VPROT_COMMITTED | VPROT_READ ); sec = (IMAGE_SECTION_HEADER *)((char *)&nt->OptionalHeader + nt->FileHeader.SizeOfOptionalHeader); for (i = 0; i < nt->FileHeader.NumberOfSections; i++) { BYTE flags = VPROT_COMMITTED; if (sec[i].Characteristics & IMAGE_SCN_MEM_EXECUTE) flags |= VPROT_EXEC; if (sec[i].Characteristics & IMAGE_SCN_MEM_READ) flags |= VPROT_READ; if (sec[i].Characteristics & IMAGE_SCN_MEM_WRITE) flags |= VPROT_WRITE; set_page_vprot( (char *)base + sec[i].VirtualAddress, sec[i].Misc.VirtualSize, flags ); } VIRTUAL_DEBUG_DUMP_VIEW( view ); if (is_beyond_limit( base, size, working_set_limit )) working_set_limit = address_space_limit; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /* set some initial values in a new TEB */ static void init_teb( TEB *teb, PEB *peb ) { struct ntdll_thread_data *thread_data = (struct ntdll_thread_data *)&teb->GdiTebBatch; teb->Peb = peb; teb->Tib.Self = &teb->Tib; teb->Tib.ExceptionList = (void *)~0ul; teb->Tib.StackBase = (void *)~0ul; teb->StaticUnicodeString.Buffer = teb->StaticUnicodeBuffer; teb->StaticUnicodeString.MaximumLength = sizeof(teb->StaticUnicodeBuffer); thread_data->request_fd = -1; thread_data->reply_fd = -1; thread_data->wait_fd[0] = -1; thread_data->wait_fd[1] = -1; list_add_head( &teb_list, &thread_data->entry ); } /*********************************************************************** * virtual_alloc_first_teb */ TEB *virtual_alloc_first_teb(void) { TEB *teb; PEB *peb; NTSTATUS status; SIZE_T data_size = page_size; SIZE_T peb_size = page_size; SIZE_T teb_size = signal_stack_mask + 1; SIZE_T total = 32 * teb_size; /* reserve space for shared user data */ status = NtAllocateVirtualMemory( NtCurrentProcess(), (void **)&user_shared_data, 0, &data_size, MEM_RESERVE | MEM_COMMIT, PAGE_READONLY ); if (status) { ERR( "wine: failed to map the shared user data: %08x\n", status ); exit(1); } NtAllocateVirtualMemory( NtCurrentProcess(), (void **)&teb_block, 0, &total, MEM_RESERVE | MEM_TOP_DOWN, PAGE_READWRITE ); teb_block_pos = 30; teb = (TEB *)((char *)teb_block + 30 * teb_size); peb = (PEB *)((char *)teb_block + 32 * teb_size - peb_size); NtAllocateVirtualMemory( NtCurrentProcess(), (void **)&teb, 0, &teb_size, MEM_COMMIT, PAGE_READWRITE ); NtAllocateVirtualMemory( NtCurrentProcess(), (void **)&peb, 0, &peb_size, MEM_COMMIT, PAGE_READWRITE ); init_teb( teb, peb ); *(ULONG_PTR *)peb->Reserved = get_image_address(); return teb; } /*********************************************************************** * virtual_alloc_teb */ NTSTATUS virtual_alloc_teb( TEB **ret_teb ) { sigset_t sigset; TEB *teb = NULL; NTSTATUS status = STATUS_SUCCESS; SIZE_T teb_size = signal_stack_mask + 1; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (next_free_teb) { teb = next_free_teb; next_free_teb = *(TEB **)teb; memset( teb, 0, sizeof(*teb) ); } else { if (!teb_block_pos) { void *addr = NULL; SIZE_T total = 32 * teb_size; if ((status = NtAllocateVirtualMemory( NtCurrentProcess(), &addr, 0, &total, MEM_RESERVE, PAGE_READWRITE ))) { server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } teb_block = addr; teb_block_pos = 32; } teb = (TEB *)((char *)teb_block + --teb_block_pos * teb_size); NtAllocateVirtualMemory( NtCurrentProcess(), (void **)&teb, 0, &teb_size, MEM_COMMIT, PAGE_READWRITE ); } init_teb( teb, NtCurrentTeb()->Peb ); *ret_teb = teb; server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if ((status = signal_alloc_thread( teb ))) { server_enter_uninterrupted_section( &virtual_mutex, &sigset ); *(TEB **)teb = next_free_teb; next_free_teb = teb; server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } return status; } /*********************************************************************** * virtual_free_teb */ void virtual_free_teb( TEB *teb ) { struct ntdll_thread_data *thread_data = (struct ntdll_thread_data *)&teb->GdiTebBatch; SIZE_T size; sigset_t sigset; signal_free_thread( teb ); if (teb->DeallocationStack) { size = 0; NtFreeVirtualMemory( GetCurrentProcess(), &teb->DeallocationStack, &size, MEM_RELEASE ); } if (thread_data->start_stack) { size = 0; NtFreeVirtualMemory( GetCurrentProcess(), &thread_data->start_stack, &size, MEM_RELEASE ); } server_enter_uninterrupted_section( &virtual_mutex, &sigset ); list_remove( &thread_data->entry ); *(TEB **)teb = next_free_teb; next_free_teb = teb; server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } /*********************************************************************** * virtual_clear_tls_index */ NTSTATUS virtual_clear_tls_index( ULONG index ) { struct ntdll_thread_data *thread_data; sigset_t sigset; if (index < TLS_MINIMUM_AVAILABLE) { server_enter_uninterrupted_section( &virtual_mutex, &sigset ); LIST_FOR_EACH_ENTRY( thread_data, &teb_list, struct ntdll_thread_data, entry ) { TEB *teb = CONTAINING_RECORD( thread_data, TEB, GdiTebBatch ); teb->TlsSlots[index] = 0; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } else { index -= TLS_MINIMUM_AVAILABLE; if (index >= 8 * sizeof(NtCurrentTeb()->Peb->TlsExpansionBitmapBits)) return STATUS_INVALID_PARAMETER; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); LIST_FOR_EACH_ENTRY( thread_data, &teb_list, struct ntdll_thread_data, entry ) { TEB *teb = CONTAINING_RECORD( thread_data, TEB, GdiTebBatch ); if (teb->TlsExpansionSlots) teb->TlsExpansionSlots[index] = 0; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } return STATUS_SUCCESS; } /*********************************************************************** * virtual_alloc_thread_stack */ NTSTATUS CDECL virtual_alloc_thread_stack( INITIAL_TEB *stack, SIZE_T reserve_size, SIZE_T commit_size, SIZE_T *pthread_size ) { struct file_view *view; NTSTATUS status; sigset_t sigset; SIZE_T size, extra_size = 0; if (!reserve_size || !commit_size) { IMAGE_NT_HEADERS *nt = RtlImageNtHeader( NtCurrentTeb()->Peb->ImageBaseAddress ); if (!reserve_size) reserve_size = nt->OptionalHeader.SizeOfStackReserve; if (!commit_size) commit_size = nt->OptionalHeader.SizeOfStackCommit; } size = max( reserve_size, commit_size ); if (size < 1024 * 1024) size = 1024 * 1024; /* Xlib needs a large stack */ size = (size + 0xffff) & ~0xffff; /* round to 64K boundary */ if (pthread_size) *pthread_size = extra_size = max( page_size, ROUND_SIZE( 0, *pthread_size )); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((status = map_view( &view, NULL, size + extra_size, FALSE, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED, 0 )) != STATUS_SUCCESS) goto done; #ifdef VALGRIND_STACK_REGISTER VALGRIND_STACK_REGISTER( view->base, (char *)view->base + view->size ); #endif /* setup no access guard page */ set_page_vprot( view->base, page_size, VPROT_COMMITTED ); set_page_vprot( (char *)view->base + page_size, page_size, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED | VPROT_GUARD ); mprotect_range( view->base, 2 * page_size, 0, 0 ); VIRTUAL_DEBUG_DUMP_VIEW( view ); if (extra_size) { struct file_view *extra_view; /* shrink the first view and create a second one for the extra size */ /* this allows the app to free the stack without freeing the thread start portion */ view->size -= extra_size; status = create_view( &extra_view, (char *)view->base + view->size, extra_size, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED ); if (status != STATUS_SUCCESS) { view->size += extra_size; delete_view( view ); goto done; } } /* note: limit is lower than base since the stack grows down */ stack->OldStackBase = 0; stack->OldStackLimit = 0; stack->DeallocationStack = view->base; stack->StackBase = (char *)view->base + view->size; stack->StackLimit = (char *)view->base + 2 * page_size; done: server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /*********************************************************************** * virtual_clear_thread_stack * * Clear the stack contents before calling the main entry point, some broken apps need that. */ void virtual_clear_thread_stack( void *stack_end ) { void *stack = NtCurrentTeb()->Tib.StackLimit; size_t size = (char *)stack_end - (char *)stack; wine_anon_mmap( stack, size, PROT_READ | PROT_WRITE, MAP_FIXED ); if (force_exec_prot) mprotect( stack, size, PROT_READ | PROT_WRITE | PROT_EXEC ); } /*********************************************************************** * virtual_map_user_shared_data */ void virtual_map_user_shared_data(void) { static const WCHAR wine_usdW[] = {'\\','K','e','r','n','e','l','O','b','j','e','c','t','s', '\\','_','_','w','i','n','e','_','u','s','e','r','_','s','h','a','r','e','d','_','d','a','t','a',0}; OBJECT_ATTRIBUTES attr = {sizeof(attr)}; UNICODE_STRING wine_usd_str; NTSTATUS status; HANDLE section; int res, fd, needs_close; RtlInitUnicodeString( &wine_usd_str, wine_usdW ); InitializeObjectAttributes( &attr, &wine_usd_str, OBJ_OPENIF, NULL, NULL ); if ((status = NtOpenSection( §ion, SECTION_ALL_ACCESS, &attr ))) { ERR( "failed to open the USD section: %08x\n", status ); exit(1); } if ((res = server_get_unix_fd( section, 0, &fd, &needs_close, NULL, NULL )) || (user_shared_data != mmap( user_shared_data, page_size, PROT_READ, MAP_SHARED|MAP_FIXED, fd, 0 ))) { ERR( "failed to remap the process USD: %d\n", res ); exit(1); } if (needs_close) close( fd ); NtClose( section ); } /*********************************************************************** * virtual_handle_fault */ NTSTATUS virtual_handle_fault( LPCVOID addr, DWORD err, BOOL on_signal_stack ) { NTSTATUS ret = STATUS_ACCESS_VIOLATION; void *page = ROUND_ADDR( addr, page_mask ); sigset_t sigset; BYTE vprot; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); vprot = get_page_vprot( page ); if (!on_signal_stack && (vprot & VPROT_GUARD)) { set_page_vprot_bits( page, page_size, 0, VPROT_GUARD ); mprotect_range( page, page_size, 0, 0 ); ret = STATUS_GUARD_PAGE_VIOLATION; } else if (err & EXCEPTION_WRITE_FAULT) { if (vprot & VPROT_WRITEWATCH) { set_page_vprot_bits( page, page_size, 0, VPROT_WRITEWATCH ); mprotect_range( page, page_size, 0, 0 ); } /* ignore fault if page is writable now */ if (get_unix_prot( get_page_vprot( page )) & PROT_WRITE) { if ((vprot & VPROT_WRITEWATCH) || is_write_watch_range( page, page_size )) ret = STATUS_SUCCESS; } } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return ret; } /*********************************************************************** * check_write_access * * Check if the memory range is writable, temporarily disabling write watches if necessary. */ static NTSTATUS check_write_access( void *base, size_t size, BOOL *has_write_watch ) { size_t i; char *addr = ROUND_ADDR( base, page_mask ); size = ROUND_SIZE( base, size ); for (i = 0; i < size; i += page_size) { BYTE vprot = get_page_vprot( addr + i ); if (vprot & VPROT_WRITEWATCH) *has_write_watch = TRUE; if (!(get_unix_prot( vprot & ~VPROT_WRITEWATCH ) & PROT_WRITE)) return STATUS_INVALID_USER_BUFFER; } if (*has_write_watch) mprotect_range( addr, size, 0, VPROT_WRITEWATCH ); /* temporarily enable write access */ return STATUS_SUCCESS; } /*********************************************************************** * virtual_locked_server_call */ unsigned int virtual_locked_server_call( void *req_ptr ) { struct __server_request_info * const req = req_ptr; sigset_t sigset; void *addr = req->reply_data; data_size_t size = req->u.req.request_header.reply_size; BOOL has_write_watch = FALSE; unsigned int ret = STATUS_ACCESS_VIOLATION; if (!size) return wine_server_call( req_ptr ); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!(ret = check_write_access( addr, size, &has_write_watch ))) { ret = server_call_unlocked( req ); if (has_write_watch) update_write_watches( addr, size, wine_server_reply_size( req )); } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return ret; } /*********************************************************************** * virtual_locked_read */ ssize_t virtual_locked_read( int fd, void *addr, size_t size ) { sigset_t sigset; BOOL has_write_watch = FALSE; int err = EFAULT; ssize_t ret = read( fd, addr, size ); if (ret != -1 || errno != EFAULT) return ret; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!check_write_access( addr, size, &has_write_watch )) { ret = read( fd, addr, size ); err = errno; if (has_write_watch) update_write_watches( addr, size, max( 0, ret )); } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); errno = err; return ret; } /*********************************************************************** * virtual_locked_pread */ ssize_t virtual_locked_pread( int fd, void *addr, size_t size, off_t offset ) { sigset_t sigset; BOOL has_write_watch = FALSE; int err = EFAULT; ssize_t ret = pread( fd, addr, size, offset ); if (ret != -1 || errno != EFAULT) return ret; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!check_write_access( addr, size, &has_write_watch )) { ret = pread( fd, addr, size, offset ); err = errno; if (has_write_watch) update_write_watches( addr, size, max( 0, ret )); } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); errno = err; return ret; } /*********************************************************************** * virtual_locked_recvmsg */ ssize_t CDECL virtual_locked_recvmsg( int fd, struct msghdr *hdr, int flags ) { sigset_t sigset; size_t i; BOOL has_write_watch = FALSE; int err = EFAULT; ssize_t ret = recvmsg( fd, hdr, flags ); if (ret != -1 || errno != EFAULT) return ret; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); for (i = 0; i < hdr->msg_iovlen; i++) if (check_write_access( hdr->msg_iov[i].iov_base, hdr->msg_iov[i].iov_len, &has_write_watch )) break; if (i == hdr->msg_iovlen) { ret = recvmsg( fd, hdr, flags ); err = errno; } if (has_write_watch) while (i--) update_write_watches( hdr->msg_iov[i].iov_base, hdr->msg_iov[i].iov_len, 0 ); server_leave_uninterrupted_section( &virtual_mutex, &sigset ); errno = err; return ret; } /*********************************************************************** * virtual_is_valid_code_address */ BOOL virtual_is_valid_code_address( const void *addr, SIZE_T size ) { struct file_view *view; BOOL ret = FALSE; sigset_t sigset; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((view = find_view( addr, size ))) ret = !(view->protect & VPROT_SYSTEM); /* system views are not visible to the app */ server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return ret; } /*********************************************************************** * virtual_handle_stack_fault * * Handle an access fault inside the current thread stack. * Return 1 if safely handled, -1 if handled into the overflow space. * Called from inside a signal handler. */ int virtual_handle_stack_fault( void *addr ) { int ret = 0; if ((char *)addr < (char *)NtCurrentTeb()->DeallocationStack) return 0; if ((char *)addr >= (char *)NtCurrentTeb()->Tib.StackBase) return 0; pthread_mutex_lock( &virtual_mutex ); /* no need for signal masking inside signal handler */ if (get_page_vprot( addr ) & VPROT_GUARD) { size_t guaranteed = max( NtCurrentTeb()->GuaranteedStackBytes, page_size * (is_win64 ? 2 : 1) ); char *page = ROUND_ADDR( addr, page_mask ); set_page_vprot_bits( page, page_size, 0, VPROT_GUARD ); mprotect_range( page, page_size, 0, 0 ); if (page >= (char *)NtCurrentTeb()->DeallocationStack + page_size + guaranteed) { set_page_vprot_bits( page - page_size, page_size, VPROT_COMMITTED | VPROT_GUARD, 0 ); mprotect_range( page - page_size, page_size, 0, 0 ); ret = 1; } else /* inside guaranteed space -> overflow exception */ { page = (char *)NtCurrentTeb()->DeallocationStack + page_size; set_page_vprot_bits( page, guaranteed, VPROT_COMMITTED, VPROT_GUARD ); mprotect_range( page, guaranteed, 0, 0 ); ret = -1; } NtCurrentTeb()->Tib.StackLimit = page; } pthread_mutex_unlock( &virtual_mutex ); return ret; } /*********************************************************************** * virtual_check_buffer_for_read * * Check if a memory buffer can be read, triggering page faults if needed for DIB section access. */ BOOL virtual_check_buffer_for_read( const void *ptr, SIZE_T size ) { if (!size) return TRUE; if (!ptr) return FALSE; __TRY { volatile const char *p = ptr; char dummy __attribute__((unused)); SIZE_T count = size; while (count > page_size) { dummy = *p; p += page_size; count -= page_size; } dummy = p[0]; dummy = p[count - 1]; } __EXCEPT_PAGE_FAULT { return FALSE; } __ENDTRY return TRUE; } /*********************************************************************** * virtual_check_buffer_for_write * * Check if a memory buffer can be written to, triggering page faults if needed for write watches. */ BOOL virtual_check_buffer_for_write( void *ptr, SIZE_T size ) { if (!size) return TRUE; if (!ptr) return FALSE; __TRY { volatile char *p = ptr; SIZE_T count = size; while (count > page_size) { *p |= 0; p += page_size; count -= page_size; } p[0] |= 0; p[count - 1] |= 0; } __EXCEPT_PAGE_FAULT { return FALSE; } __ENDTRY return TRUE; } /************************************************************* * IsBadStringPtrA * * IsBadStringPtrA replacement for ntdll, to catch exception in debug traces. */ BOOL WINAPI IsBadStringPtrA( LPCSTR str, UINT_PTR max ) { if (!str) return TRUE; __TRY { volatile const char *p = str; while (p != str + max) if (!*p++) break; } __EXCEPT_PAGE_FAULT { return TRUE; } __ENDTRY return FALSE; } /************************************************************* * IsBadStringPtrW * * IsBadStringPtrW replacement for ntdll, to catch exception in debug traces. */ BOOL WINAPI IsBadStringPtrW( LPCWSTR str, UINT_PTR max ) { if (!str) return TRUE; __TRY { volatile const WCHAR *p = str; while (p != str + max) if (!*p++) break; } __EXCEPT_PAGE_FAULT { return TRUE; } __ENDTRY return FALSE; } /*********************************************************************** * virtual_uninterrupted_read_memory * * Similar to NtReadVirtualMemory, but without wineserver calls. Moreover * permissions are checked before accessing each page, to ensure that no * exceptions can happen. */ SIZE_T virtual_uninterrupted_read_memory( const void *addr, void *buffer, SIZE_T size ) { struct file_view *view; sigset_t sigset; SIZE_T bytes_read = 0; if (!size) return 0; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((view = find_view( addr, size ))) { if (!(view->protect & VPROT_SYSTEM)) { while (bytes_read < size && (get_unix_prot( get_page_vprot( addr )) & PROT_READ)) { SIZE_T block_size = min( size - bytes_read, page_size - ((UINT_PTR)addr & page_mask) ); memcpy( buffer, addr, block_size ); addr = (const void *)((const char *)addr + block_size); buffer = (void *)((char *)buffer + block_size); bytes_read += block_size; } } } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return bytes_read; } /*********************************************************************** * virtual_uninterrupted_write_memory * * Similar to NtWriteVirtualMemory, but without wineserver calls. Moreover * permissions are checked before accessing each page, to ensure that no * exceptions can happen. */ NTSTATUS virtual_uninterrupted_write_memory( void *addr, const void *buffer, SIZE_T size ) { BOOL has_write_watch = FALSE; sigset_t sigset; NTSTATUS ret; if (!size) return STATUS_SUCCESS; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!(ret = check_write_access( addr, size, &has_write_watch ))) { memcpy( addr, buffer, size ); if (has_write_watch) update_write_watches( addr, size, size ); } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return ret; } /*********************************************************************** * virtual_set_force_exec * * Whether to force exec prot on all views. */ void virtual_set_force_exec( BOOL enable ) { struct file_view *view; sigset_t sigset; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!force_exec_prot != !enable) /* change all existing views */ { force_exec_prot = enable; WINE_RB_FOR_EACH_ENTRY( view, &views_tree, struct file_view, entry ) { /* file mappings are always accessible */ BYTE commit = is_view_valloc( view ) ? 0 : VPROT_COMMITTED; mprotect_range( view->base, view->size, commit, 0 ); } } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } struct free_range { char *base; char *limit; }; /* free reserved areas above the limit; callback for mmap_enum_reserved_areas */ static int CDECL free_reserved_memory( void *base, SIZE_T size, void *arg ) { struct free_range *range = arg; if ((char *)base >= range->limit) return 0; if ((char *)base + size <= range->base) return 0; if ((char *)base < range->base) { size -= range->base - (char *)base; base = range->base; } if ((char *)base + size > range->limit) size = range->limit - (char *)base; remove_reserved_area( base, size ); return 1; /* stop enumeration since the list has changed */ } /*********************************************************************** * virtual_release_address_space * * Release some address space once we have loaded and initialized the app. */ void CDECL virtual_release_address_space(void) { struct free_range range; sigset_t sigset; if (is_win64) return; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); range.base = (char *)0x82000000; range.limit = user_space_limit; if (range.limit > range.base) { while (mmap_enum_reserved_areas( free_reserved_memory, &range, 1 )) /* nothing */; #ifdef __APPLE__ /* On macOS, we still want to free some of low memory, for OpenGL resources */ range.base = (char *)0x40000000; #else range.base = NULL; #endif } else range.base = (char *)0x20000000; if (range.base) { range.limit = (char *)0x7f000000; while (mmap_enum_reserved_areas( free_reserved_memory, &range, 0 )) /* nothing */; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); } /*********************************************************************** * virtual_set_large_address_space * * Enable use of a large address space when allowed by the application. */ void CDECL virtual_set_large_address_space(void) { IMAGE_NT_HEADERS *nt = RtlImageNtHeader( NtCurrentTeb()->Peb->ImageBaseAddress ); if (!(nt->FileHeader.Characteristics & IMAGE_FILE_LARGE_ADDRESS_AWARE)) return; /* no large address space on win9x */ if (NtCurrentTeb()->Peb->OSPlatformId != VER_PLATFORM_WIN32_NT) return; user_space_limit = working_set_limit = address_space_limit; } /*********************************************************************** * NtAllocateVirtualMemory (NTDLL.@) * ZwAllocateVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtAllocateVirtualMemory( HANDLE process, PVOID *ret, ULONG_PTR zero_bits, SIZE_T *size_ptr, ULONG type, ULONG protect ) { void *base; unsigned int vprot; BOOL is_dos_memory = FALSE; struct file_view *view; sigset_t sigset; SIZE_T size = *size_ptr; NTSTATUS status = STATUS_SUCCESS; unsigned short zero_bits_64 = zero_bits_win_to_64( zero_bits ); TRACE("%p %p %08lx %x %08x\n", process, *ret, size, type, protect ); if (!size) return STATUS_INVALID_PARAMETER; if (zero_bits > 21 && zero_bits < 32) return STATUS_INVALID_PARAMETER_3; if (!is_win64 && !is_wow64 && zero_bits >= 32) return STATUS_INVALID_PARAMETER_3; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_alloc.type = APC_VIRTUAL_ALLOC; call.virtual_alloc.addr = wine_server_client_ptr( *ret ); call.virtual_alloc.size = *size_ptr; call.virtual_alloc.zero_bits = zero_bits; call.virtual_alloc.op_type = type; call.virtual_alloc.prot = protect; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_alloc.status == STATUS_SUCCESS) { *ret = wine_server_get_ptr( result.virtual_alloc.addr ); *size_ptr = result.virtual_alloc.size; } return result.virtual_alloc.status; } /* Round parameters to a page boundary */ if (is_beyond_limit( 0, size, working_set_limit )) return STATUS_WORKING_SET_LIMIT_RANGE; if (*ret) { if (type & MEM_RESERVE) /* Round down to 64k boundary */ base = ROUND_ADDR( *ret, granularity_mask ); else base = ROUND_ADDR( *ret, page_mask ); size = (((UINT_PTR)*ret + size + page_mask) & ~page_mask) - (UINT_PTR)base; /* disallow low 64k, wrap-around and kernel space */ if (((char *)base < (char *)0x10000) || ((char *)base + size < (char *)base) || is_beyond_limit( base, size, address_space_limit )) { /* address 1 is magic to mean DOS area */ if (!base && *ret == (void *)1 && size == 0x110000) is_dos_memory = TRUE; else return STATUS_INVALID_PARAMETER; } } else { base = NULL; size = (size + page_mask) & ~page_mask; } /* Compute the alloc type flags */ if (!(type & (MEM_COMMIT | MEM_RESERVE | MEM_RESET)) || (type & ~(MEM_COMMIT | MEM_RESERVE | MEM_TOP_DOWN | MEM_WRITE_WATCH | MEM_RESET))) { WARN("called with wrong alloc type flags (%08x) !\n", type); return STATUS_INVALID_PARAMETER; } /* Reserve the memory */ server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((type & MEM_RESERVE) || !base) { if (!(status = get_vprot_flags( protect, &vprot, FALSE ))) { if (type & MEM_COMMIT) vprot |= VPROT_COMMITTED; if (type & MEM_WRITE_WATCH) vprot |= VPROT_WRITEWATCH; if (protect & PAGE_NOCACHE) vprot |= SEC_NOCACHE; if (vprot & VPROT_WRITECOPY) status = STATUS_INVALID_PAGE_PROTECTION; else if (is_dos_memory) status = allocate_dos_memory( &view, vprot ); else status = map_view( &view, base, size, type & MEM_TOP_DOWN, vprot, zero_bits_64 ); if (status == STATUS_SUCCESS) base = view->base; } } else if (type & MEM_RESET) { if (!(view = find_view( base, size ))) status = STATUS_NOT_MAPPED_VIEW; else madvise( base, size, MADV_DONTNEED ); } else /* commit the pages */ { if (!(view = find_view( base, size ))) status = STATUS_NOT_MAPPED_VIEW; else if (view->protect & SEC_FILE) status = STATUS_ALREADY_COMMITTED; else if (!(status = set_protection( view, base, size, protect )) && (view->protect & SEC_RESERVE)) { SERVER_START_REQ( add_mapping_committed_range ) { req->base = wine_server_client_ptr( view->base ); req->offset = (char *)base - (char *)view->base; req->size = size; wine_server_call( req ); } SERVER_END_REQ; } } if (!status) VIRTUAL_DEBUG_DUMP_VIEW( view ); server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if (status == STATUS_SUCCESS) { *ret = base; *size_ptr = size; } return status; } /*********************************************************************** * NtFreeVirtualMemory (NTDLL.@) * ZwFreeVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtFreeVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr, ULONG type ) { struct file_view *view; char *base; sigset_t sigset; NTSTATUS status = STATUS_SUCCESS; LPVOID addr = *addr_ptr; SIZE_T size = *size_ptr; TRACE("%p %p %08lx %x\n", process, addr, size, type ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_free.type = APC_VIRTUAL_FREE; call.virtual_free.addr = wine_server_client_ptr( addr ); call.virtual_free.size = size; call.virtual_free.op_type = type; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_free.status == STATUS_SUCCESS) { *addr_ptr = wine_server_get_ptr( result.virtual_free.addr ); *size_ptr = result.virtual_free.size; } return result.virtual_free.status; } /* Fix the parameters */ size = ROUND_SIZE( addr, size ); base = ROUND_ADDR( addr, page_mask ); /* avoid freeing the DOS area when a broken app passes a NULL pointer */ if (!base) return STATUS_INVALID_PARAMETER; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!(view = find_view( base, size )) || !is_view_valloc( view )) { status = STATUS_INVALID_PARAMETER; } else if (type == MEM_RELEASE) { /* Free the pages */ if (size || (base != view->base)) status = STATUS_INVALID_PARAMETER; else { delete_view( view ); *addr_ptr = base; *size_ptr = size; } } else if (type == MEM_DECOMMIT) { status = decommit_pages( view, base - (char *)view->base, size ); if (status == STATUS_SUCCESS) { *addr_ptr = base; *size_ptr = size; } } else { WARN("called with wrong free type flags (%08x) !\n", type); status = STATUS_INVALID_PARAMETER; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /*********************************************************************** * NtProtectVirtualMemory (NTDLL.@) * ZwProtectVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtProtectVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr, ULONG new_prot, ULONG *old_prot ) { struct file_view *view; sigset_t sigset; NTSTATUS status = STATUS_SUCCESS; char *base; BYTE vprot; SIZE_T size = *size_ptr; LPVOID addr = *addr_ptr; DWORD old; TRACE("%p %p %08lx %08x\n", process, addr, size, new_prot ); if (!old_prot) return STATUS_ACCESS_VIOLATION; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_protect.type = APC_VIRTUAL_PROTECT; call.virtual_protect.addr = wine_server_client_ptr( addr ); call.virtual_protect.size = size; call.virtual_protect.prot = new_prot; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_protect.status == STATUS_SUCCESS) { *addr_ptr = wine_server_get_ptr( result.virtual_protect.addr ); *size_ptr = result.virtual_protect.size; *old_prot = result.virtual_protect.prot; } return result.virtual_protect.status; } /* Fix the parameters */ size = ROUND_SIZE( addr, size ); base = ROUND_ADDR( addr, page_mask ); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((view = find_view( base, size ))) { /* Make sure all the pages are committed */ if (get_committed_size( view, base, &vprot ) >= size && (vprot & VPROT_COMMITTED)) { old = get_win32_prot( vprot, view->protect ); status = set_protection( view, base, size, new_prot ); } else status = STATUS_NOT_COMMITTED; } else status = STATUS_INVALID_PARAMETER; if (!status) VIRTUAL_DEBUG_DUMP_VIEW( view ); server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if (status == STATUS_SUCCESS) { *addr_ptr = base; *size_ptr = size; *old_prot = old; } return status; } /* retrieve state for a free memory area; callback for mmap_enum_reserved_areas */ static int CDECL get_free_mem_state_callback( void *start, SIZE_T size, void *arg ) { MEMORY_BASIC_INFORMATION *info = arg; void *end = (char *)start + size; if ((char *)info->BaseAddress + info->RegionSize <= (char *)start) return 0; if (info->BaseAddress >= end) { if (info->AllocationBase < end) info->AllocationBase = end; return 0; } if (info->BaseAddress >= start || start <= address_space_start) { /* it's a real free area */ info->State = MEM_FREE; info->Protect = PAGE_NOACCESS; info->AllocationBase = 0; info->AllocationProtect = 0; info->Type = 0; if ((char *)info->BaseAddress + info->RegionSize > (char *)end) info->RegionSize = (char *)end - (char *)info->BaseAddress; } else /* outside of the reserved area, pretend it's allocated */ { info->RegionSize = (char *)start - (char *)info->BaseAddress; info->State = MEM_RESERVE; info->Protect = PAGE_NOACCESS; info->AllocationProtect = PAGE_NOACCESS; info->Type = MEM_PRIVATE; } return 1; } /* get basic information about a memory block */ static NTSTATUS get_basic_memory_info( HANDLE process, LPCVOID addr, MEMORY_BASIC_INFORMATION *info, SIZE_T len, SIZE_T *res_len ) { struct file_view *view; char *base, *alloc_base = 0, *alloc_end = working_set_limit; struct wine_rb_entry *ptr; sigset_t sigset; if (len < sizeof(MEMORY_BASIC_INFORMATION)) return STATUS_INFO_LENGTH_MISMATCH; if (process != NtCurrentProcess()) { NTSTATUS status; apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_query.type = APC_VIRTUAL_QUERY; call.virtual_query.addr = wine_server_client_ptr( addr ); status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_query.status == STATUS_SUCCESS) { info->BaseAddress = wine_server_get_ptr( result.virtual_query.base ); info->AllocationBase = wine_server_get_ptr( result.virtual_query.alloc_base ); info->RegionSize = result.virtual_query.size; info->Protect = result.virtual_query.prot; info->AllocationProtect = result.virtual_query.alloc_prot; info->State = (DWORD)result.virtual_query.state << 12; info->Type = (DWORD)result.virtual_query.alloc_type << 16; if (info->RegionSize != result.virtual_query.size) /* truncated */ return STATUS_INVALID_PARAMETER; /* FIXME */ if (res_len) *res_len = sizeof(*info); } return result.virtual_query.status; } base = ROUND_ADDR( addr, page_mask ); if (is_beyond_limit( base, 1, working_set_limit )) return STATUS_INVALID_PARAMETER; /* Find the view containing the address */ server_enter_uninterrupted_section( &virtual_mutex, &sigset ); ptr = views_tree.root; while (ptr) { view = WINE_RB_ENTRY_VALUE( ptr, struct file_view, entry ); if ((char *)view->base > base) { alloc_end = view->base; ptr = ptr->left; } else if ((char *)view->base + view->size <= base) { alloc_base = (char *)view->base + view->size; ptr = ptr->right; } else { alloc_base = view->base; alloc_end = (char *)view->base + view->size; break; } } /* Fill the info structure */ info->AllocationBase = alloc_base; info->BaseAddress = base; info->RegionSize = alloc_end - base; if (!ptr) { if (!mmap_enum_reserved_areas( get_free_mem_state_callback, info, 0 )) { /* not in a reserved area at all, pretend it's allocated */ #ifdef __i386__ if (base >= (char *)address_space_start) { info->State = MEM_RESERVE; info->Protect = PAGE_NOACCESS; info->AllocationProtect = PAGE_NOACCESS; info->Type = MEM_PRIVATE; } else #endif { info->State = MEM_FREE; info->Protect = PAGE_NOACCESS; info->AllocationBase = 0; info->AllocationProtect = 0; info->Type = 0; } } } else { BYTE vprot; char *ptr; SIZE_T range_size = get_committed_size( view, base, &vprot ); info->State = (vprot & VPROT_COMMITTED) ? MEM_COMMIT : MEM_RESERVE; info->Protect = (vprot & VPROT_COMMITTED) ? get_win32_prot( vprot, view->protect ) : 0; info->AllocationProtect = get_win32_prot( view->protect, view->protect ); if (view->protect & SEC_IMAGE) info->Type = MEM_IMAGE; else if (view->protect & (SEC_FILE | SEC_RESERVE | SEC_COMMIT)) info->Type = MEM_MAPPED; else info->Type = MEM_PRIVATE; for (ptr = base; ptr < base + range_size; ptr += page_size) if ((get_page_vprot( ptr ) ^ vprot) & ~VPROT_WRITEWATCH) break; info->RegionSize = ptr - base; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if (res_len) *res_len = sizeof(*info); return STATUS_SUCCESS; } static NTSTATUS get_working_set_ex( HANDLE process, LPCVOID addr, MEMORY_WORKING_SET_EX_INFORMATION *info, SIZE_T len, SIZE_T *res_len ) { FILE *f; MEMORY_WORKING_SET_EX_INFORMATION *p; sigset_t sigset; if (process != NtCurrentProcess()) { FIXME( "(process=%p,addr=%p) Unimplemented information class: MemoryWorkingSetExInformation\n", process, addr ); return STATUS_INVALID_INFO_CLASS; } f = fopen( "/proc/self/pagemap", "rb" ); if (!f) { static int once; if (!once++) WARN( "unable to open /proc/self/pagemap\n" ); } server_enter_uninterrupted_section( &virtual_mutex, &sigset ); for (p = info; (UINT_PTR)(p + 1) <= (UINT_PTR)info + len; p++) { BYTE vprot; UINT64 pagemap; struct file_view *view; memset( &p->VirtualAttributes, 0, sizeof(p->VirtualAttributes) ); /* If we don't have pagemap information, default to invalid. */ if (!f || fseek( f, ((UINT_PTR)p->VirtualAddress >> 12) * sizeof(pagemap), SEEK_SET ) == -1 || fread( &pagemap, sizeof(pagemap), 1, f ) != 1) { pagemap = 0; } if ((view = find_view( p->VirtualAddress, 0 )) && get_committed_size( view, p->VirtualAddress, &vprot ) && (vprot & VPROT_COMMITTED)) { p->VirtualAttributes.Valid = !(vprot & VPROT_GUARD) && (vprot & 0x0f) && (pagemap >> 63); p->VirtualAttributes.Shared = !is_view_valloc( view ) && ((pagemap >> 61) & 1); if (p->VirtualAttributes.Shared && p->VirtualAttributes.Valid) p->VirtualAttributes.ShareCount = 1; /* FIXME */ if (p->VirtualAttributes.Valid) p->VirtualAttributes.Win32Protection = get_win32_prot( vprot, view->protect ); } } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); if (f) fclose( f ); if (res_len) *res_len = (UINT_PTR)p - (UINT_PTR)info; return STATUS_SUCCESS; } #define UNIMPLEMENTED_INFO_CLASS(c) \ case c: \ FIXME("(process=%p,addr=%p) Unimplemented information class: " #c "\n", process, addr); \ return STATUS_INVALID_INFO_CLASS /*********************************************************************** * NtQueryVirtualMemory (NTDLL.@) * ZwQueryVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtQueryVirtualMemory( HANDLE process, LPCVOID addr, MEMORY_INFORMATION_CLASS info_class, PVOID buffer, SIZE_T len, SIZE_T *res_len ) { TRACE("(%p, %p, info_class=%d, %p, %ld, %p)\n", process, addr, info_class, buffer, len, res_len); switch(info_class) { case MemoryBasicInformation: return get_basic_memory_info( process, addr, buffer, len, res_len ); case MemoryWorkingSetExInformation: return get_working_set_ex( process, addr, buffer, len, res_len ); UNIMPLEMENTED_INFO_CLASS(MemoryWorkingSetList); UNIMPLEMENTED_INFO_CLASS(MemorySectionName); UNIMPLEMENTED_INFO_CLASS(MemoryBasicVlmInformation); default: FIXME("(%p,%p,info_class=%d,%p,%ld,%p) Unknown information class\n", process, addr, info_class, buffer, len, res_len); return STATUS_INVALID_INFO_CLASS; } } /*********************************************************************** * NtLockVirtualMemory (NTDLL.@) * ZwLockVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtLockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown ) { NTSTATUS status = STATUS_SUCCESS; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_lock.type = APC_VIRTUAL_LOCK; call.virtual_lock.addr = wine_server_client_ptr( *addr ); call.virtual_lock.size = *size; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_lock.status == STATUS_SUCCESS) { *addr = wine_server_get_ptr( result.virtual_lock.addr ); *size = result.virtual_lock.size; } return result.virtual_lock.status; } *size = ROUND_SIZE( *addr, *size ); *addr = ROUND_ADDR( *addr, page_mask ); if (mlock( *addr, *size )) status = STATUS_ACCESS_DENIED; return status; } /*********************************************************************** * NtUnlockVirtualMemory (NTDLL.@) * ZwUnlockVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtUnlockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown ) { NTSTATUS status = STATUS_SUCCESS; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_unlock.type = APC_VIRTUAL_UNLOCK; call.virtual_unlock.addr = wine_server_client_ptr( *addr ); call.virtual_unlock.size = *size; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_unlock.status == STATUS_SUCCESS) { *addr = wine_server_get_ptr( result.virtual_unlock.addr ); *size = result.virtual_unlock.size; } return result.virtual_unlock.status; } *size = ROUND_SIZE( *addr, *size ); *addr = ROUND_ADDR( *addr, page_mask ); if (munlock( *addr, *size )) status = STATUS_ACCESS_DENIED; return status; } /*********************************************************************** * NtMapViewOfSection (NTDLL.@) * ZwMapViewOfSection (NTDLL.@) */ NTSTATUS WINAPI NtMapViewOfSection( HANDLE handle, HANDLE process, PVOID *addr_ptr, ULONG_PTR zero_bits, SIZE_T commit_size, const LARGE_INTEGER *offset_ptr, SIZE_T *size_ptr, SECTION_INHERIT inherit, ULONG alloc_type, ULONG protect ) { NTSTATUS res; SIZE_T mask = granularity_mask; pe_image_info_t image_info; LARGE_INTEGER offset; unsigned short zero_bits_64 = zero_bits_win_to_64( zero_bits ); offset.QuadPart = offset_ptr ? offset_ptr->QuadPart : 0; TRACE("handle=%p process=%p addr=%p off=%x%08x size=%lx access=%x\n", handle, process, *addr_ptr, offset.u.HighPart, offset.u.LowPart, *size_ptr, protect ); /* Check parameters */ if (zero_bits > 21 && zero_bits < 32) return STATUS_INVALID_PARAMETER_4; if (!is_win64 && !is_wow64 && zero_bits >= 32) return STATUS_INVALID_PARAMETER_4; /* If both addr_ptr and zero_bits are passed, they have match */ if (*addr_ptr && zero_bits && zero_bits < 32 && (((UINT_PTR)*addr_ptr) >> (32 - zero_bits))) return STATUS_INVALID_PARAMETER_4; if (*addr_ptr && zero_bits >= 32 && (((UINT_PTR)*addr_ptr) & ~zero_bits)) return STATUS_INVALID_PARAMETER_4; #ifndef _WIN64 if (!is_wow64 && (alloc_type & AT_ROUND_TO_PAGE)) { *addr_ptr = ROUND_ADDR( *addr_ptr, page_mask ); mask = page_mask; } #endif if ((offset.u.LowPart & mask) || (*addr_ptr && ((UINT_PTR)*addr_ptr & mask))) return STATUS_MAPPED_ALIGNMENT; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.map_view.type = APC_MAP_VIEW; call.map_view.handle = wine_server_obj_handle( handle ); call.map_view.addr = wine_server_client_ptr( *addr_ptr ); call.map_view.size = *size_ptr; call.map_view.offset = offset.QuadPart; call.map_view.zero_bits = zero_bits; call.map_view.alloc_type = alloc_type; call.map_view.prot = protect; res = server_queue_process_apc( process, &call, &result ); if (res != STATUS_SUCCESS) return res; if ((NTSTATUS)result.map_view.status >= 0) { *addr_ptr = wine_server_get_ptr( result.map_view.addr ); *size_ptr = result.map_view.size; } return result.map_view.status; } return virtual_map_section( handle, addr_ptr, zero_bits_64, commit_size, offset_ptr, size_ptr, alloc_type, protect, &image_info ); } /*********************************************************************** * NtUnmapViewOfSection (NTDLL.@) * ZwUnmapViewOfSection (NTDLL.@) */ NTSTATUS WINAPI NtUnmapViewOfSection( HANDLE process, PVOID addr ) { struct file_view *view; NTSTATUS status = STATUS_NOT_MAPPED_VIEW; sigset_t sigset; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.unmap_view.type = APC_UNMAP_VIEW; call.unmap_view.addr = wine_server_client_ptr( addr ); status = server_queue_process_apc( process, &call, &result ); if (status == STATUS_SUCCESS) status = result.unmap_view.status; return status; } server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if ((view = find_view( addr, 0 )) && !is_view_valloc( view )) { if (!(view->protect & VPROT_SYSTEM)) { SERVER_START_REQ( unmap_view ) { req->base = wine_server_client_ptr( view->base ); status = wine_server_call( req ); } SERVER_END_REQ; if (!status) delete_view( view ); else FIXME( "failed to unmap %p %x\n", view->base, status ); } else { delete_view( view ); status = STATUS_SUCCESS; } } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /****************************************************************************** * virtual_fill_image_information * * Helper for NtQuerySection. */ void virtual_fill_image_information( const pe_image_info_t *pe_info, SECTION_IMAGE_INFORMATION *info ) { info->TransferAddress = wine_server_get_ptr( pe_info->entry_point ); info->ZeroBits = pe_info->zerobits; info->MaximumStackSize = pe_info->stack_size; info->CommittedStackSize = pe_info->stack_commit; info->SubSystemType = pe_info->subsystem; info->SubsystemVersionLow = pe_info->subsystem_low; info->SubsystemVersionHigh = pe_info->subsystem_high; info->GpValue = pe_info->gp; info->ImageCharacteristics = pe_info->image_charact; info->DllCharacteristics = pe_info->dll_charact; info->Machine = pe_info->machine; info->ImageContainsCode = pe_info->contains_code; info->ImageFlags = pe_info->image_flags & ~(IMAGE_FLAGS_WineBuiltin|IMAGE_FLAGS_WineFakeDll); info->LoaderFlags = pe_info->loader_flags; info->ImageFileSize = pe_info->file_size; info->CheckSum = pe_info->checksum; #ifndef _WIN64 /* don't return 64-bit values to 32-bit processes */ if (pe_info->machine == IMAGE_FILE_MACHINE_AMD64 || pe_info->machine == IMAGE_FILE_MACHINE_ARM64) { info->TransferAddress = (void *)0x81231234; /* sic */ info->MaximumStackSize = 0x100000; info->CommittedStackSize = 0x10000; } #endif } /****************************************************************************** * NtQuerySection (NTDLL.@) * ZwQuerySection (NTDLL.@) */ NTSTATUS WINAPI NtQuerySection( HANDLE handle, SECTION_INFORMATION_CLASS class, void *ptr, SIZE_T size, SIZE_T *ret_size ) { NTSTATUS status; pe_image_info_t image_info; switch (class) { case SectionBasicInformation: if (size < sizeof(SECTION_BASIC_INFORMATION)) return STATUS_INFO_LENGTH_MISMATCH; break; case SectionImageInformation: if (size < sizeof(SECTION_IMAGE_INFORMATION)) return STATUS_INFO_LENGTH_MISMATCH; break; default: FIXME( "class %u not implemented\n", class ); return STATUS_NOT_IMPLEMENTED; } if (!ptr) return STATUS_ACCESS_VIOLATION; SERVER_START_REQ( get_mapping_info ) { req->handle = wine_server_obj_handle( handle ); req->access = SECTION_QUERY; wine_server_set_reply( req, &image_info, sizeof(image_info) ); if (!(status = wine_server_call( req ))) { if (class == SectionBasicInformation) { SECTION_BASIC_INFORMATION *info = ptr; info->Attributes = reply->flags; info->BaseAddress = NULL; info->Size.QuadPart = reply->size; if (ret_size) *ret_size = sizeof(*info); } else if (reply->flags & SEC_IMAGE) { SECTION_IMAGE_INFORMATION *info = ptr; virtual_fill_image_information( &image_info, info ); if (ret_size) *ret_size = sizeof(*info); } else status = STATUS_SECTION_NOT_IMAGE; } } SERVER_END_REQ; return status; } /*********************************************************************** * NtFlushVirtualMemory (NTDLL.@) * ZwFlushVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtFlushVirtualMemory( HANDLE process, LPCVOID *addr_ptr, SIZE_T *size_ptr, ULONG unknown ) { struct file_view *view; NTSTATUS status = STATUS_SUCCESS; sigset_t sigset; void *addr = ROUND_ADDR( *addr_ptr, page_mask ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_flush.type = APC_VIRTUAL_FLUSH; call.virtual_flush.addr = wine_server_client_ptr( addr ); call.virtual_flush.size = *size_ptr; status = server_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_flush.status == STATUS_SUCCESS) { *addr_ptr = wine_server_get_ptr( result.virtual_flush.addr ); *size_ptr = result.virtual_flush.size; } return result.virtual_flush.status; } server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (!(view = find_view( addr, *size_ptr ))) status = STATUS_INVALID_PARAMETER; else { if (!*size_ptr) *size_ptr = view->size; *addr_ptr = addr; #ifdef MS_ASYNC if (msync( addr, *size_ptr, MS_ASYNC )) status = STATUS_NOT_MAPPED_DATA; #endif } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /*********************************************************************** * NtGetWriteWatch (NTDLL.@) * ZwGetWriteWatch (NTDLL.@) */ NTSTATUS WINAPI NtGetWriteWatch( HANDLE process, ULONG flags, PVOID base, SIZE_T size, PVOID *addresses, ULONG_PTR *count, ULONG *granularity ) { NTSTATUS status = STATUS_SUCCESS; sigset_t sigset; size = ROUND_SIZE( base, size ); base = ROUND_ADDR( base, page_mask ); if (!count || !granularity) return STATUS_ACCESS_VIOLATION; if (!*count || !size) return STATUS_INVALID_PARAMETER; if (flags & ~WRITE_WATCH_FLAG_RESET) return STATUS_INVALID_PARAMETER; if (!addresses) return STATUS_ACCESS_VIOLATION; TRACE( "%p %x %p-%p %p %lu\n", process, flags, base, (char *)base + size, addresses, *count ); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (is_write_watch_range( base, size )) { ULONG_PTR pos = 0; char *addr = base; char *end = addr + size; while (pos < *count && addr < end) { if (!(get_page_vprot( addr ) & VPROT_WRITEWATCH)) addresses[pos++] = addr; addr += page_size; } if (flags & WRITE_WATCH_FLAG_RESET) reset_write_watches( base, addr - (char *)base ); *count = pos; *granularity = page_size; } else status = STATUS_INVALID_PARAMETER; server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /*********************************************************************** * NtResetWriteWatch (NTDLL.@) * ZwResetWriteWatch (NTDLL.@) */ NTSTATUS WINAPI NtResetWriteWatch( HANDLE process, PVOID base, SIZE_T size ) { NTSTATUS status = STATUS_SUCCESS; sigset_t sigset; size = ROUND_SIZE( base, size ); base = ROUND_ADDR( base, page_mask ); TRACE( "%p %p-%p\n", process, base, (char *)base + size ); if (!size) return STATUS_INVALID_PARAMETER; server_enter_uninterrupted_section( &virtual_mutex, &sigset ); if (is_write_watch_range( base, size )) reset_write_watches( base, size ); else status = STATUS_INVALID_PARAMETER; server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /*********************************************************************** * NtReadVirtualMemory (NTDLL.@) * ZwReadVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtReadVirtualMemory( HANDLE process, const void *addr, void *buffer, SIZE_T size, SIZE_T *bytes_read ) { NTSTATUS status; if (virtual_check_buffer_for_write( buffer, size )) { SERVER_START_REQ( read_process_memory ) { req->handle = wine_server_obj_handle( process ); req->addr = wine_server_client_ptr( addr ); wine_server_set_reply( req, buffer, size ); if ((status = wine_server_call( req ))) size = 0; } SERVER_END_REQ; } else { status = STATUS_ACCESS_VIOLATION; size = 0; } if (bytes_read) *bytes_read = size; return status; } /*********************************************************************** * NtWriteVirtualMemory (NTDLL.@) * ZwWriteVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtWriteVirtualMemory( HANDLE process, void *addr, const void *buffer, SIZE_T size, SIZE_T *bytes_written ) { NTSTATUS status; if (virtual_check_buffer_for_read( buffer, size )) { SERVER_START_REQ( write_process_memory ) { req->handle = wine_server_obj_handle( process ); req->addr = wine_server_client_ptr( addr ); wine_server_add_data( req, buffer, size ); if ((status = wine_server_call( req ))) size = 0; } SERVER_END_REQ; } else { status = STATUS_PARTIAL_COPY; size = 0; } if (bytes_written) *bytes_written = size; return status; } /*********************************************************************** * NtAreMappedFilesTheSame (NTDLL.@) * ZwAreMappedFilesTheSame (NTDLL.@) */ NTSTATUS WINAPI NtAreMappedFilesTheSame(PVOID addr1, PVOID addr2) { struct file_view *view1, *view2; NTSTATUS status; sigset_t sigset; TRACE("%p %p\n", addr1, addr2); server_enter_uninterrupted_section( &virtual_mutex, &sigset ); view1 = find_view( addr1, 0 ); view2 = find_view( addr2, 0 ); if (!view1 || !view2) status = STATUS_INVALID_ADDRESS; else if (is_view_valloc( view1 ) || is_view_valloc( view2 )) status = STATUS_CONFLICTING_ADDRESSES; else if (view1 == view2) status = STATUS_SUCCESS; else if ((view1->protect & VPROT_SYSTEM) || (view2->protect & VPROT_SYSTEM)) status = STATUS_NOT_SAME_DEVICE; else { SERVER_START_REQ( is_same_mapping ) { req->base1 = wine_server_client_ptr( view1->base ); req->base2 = wine_server_client_ptr( view2->base ); status = wine_server_call( req ); } SERVER_END_REQ; } server_leave_uninterrupted_section( &virtual_mutex, &sigset ); return status; } /********************************************************************** * NtFlushInstructionCache (NTDLL.@) */ NTSTATUS WINAPI NtFlushInstructionCache( HANDLE handle, const void *addr, SIZE_T size ) { #if defined(__x86_64__) || defined(__i386__) /* no-op */ #elif defined(HAVE___CLEAR_CACHE) if (handle == GetCurrentProcess()) { __clear_cache( (char *)addr, (char *)addr + size ); } else { static int once; if (!once++) FIXME( "%p %p %ld other process not supported\n", handle, addr, size ); } #else static int once; if (!once++) FIXME( "%p %p %ld\n", handle, addr, size ); #endif return STATUS_SUCCESS; } /********************************************************************** * NtFlushProcessWriteBuffers (NTDLL.@) */ void WINAPI NtFlushProcessWriteBuffers(void) { static int once = 0; if (!once++) FIXME( "stub\n" ); }