/* * Win32 virtual memory functions * * Copyright 1997, 2002 Alexandre Julliard * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA */ #include "config.h" #include "wine/port.h" #include #include #ifdef HAVE_SYS_ERRNO_H #include #endif #include #ifdef HAVE_UNISTD_H # include #endif #include #include #include #include #include #ifdef HAVE_SYS_STAT_H # include #endif #ifdef HAVE_SYS_MMAN_H # include #endif #ifdef HAVE_VALGRIND_VALGRIND_H # include #endif #define NONAMELESSUNION #define NONAMELESSSTRUCT #include "ntstatus.h" #define WIN32_NO_STATUS #include "windef.h" #include "winternl.h" #include "wine/library.h" #include "wine/server.h" #include "wine/list.h" #include "wine/debug.h" #include "ntdll_misc.h" WINE_DEFAULT_DEBUG_CHANNEL(virtual); WINE_DECLARE_DEBUG_CHANNEL(module); #ifndef MS_SYNC #define MS_SYNC 0 #endif #ifndef MAP_NORESERVE #define MAP_NORESERVE 0 #endif /* File view */ typedef struct file_view { struct list entry; /* Entry in global view list */ void *base; /* Base address */ size_t size; /* Size in bytes */ HANDLE mapping; /* Handle to the file mapping */ unsigned int protect; /* Protection for all pages at allocation time */ BYTE prot[1]; /* Protection byte for each page */ } FILE_VIEW; /* Conversion from VPROT_* to Win32 flags */ static const BYTE VIRTUAL_Win32Flags[16] = { PAGE_NOACCESS, /* 0 */ PAGE_READONLY, /* READ */ PAGE_READWRITE, /* WRITE */ PAGE_READWRITE, /* READ | WRITE */ PAGE_EXECUTE, /* EXEC */ PAGE_EXECUTE_READ, /* READ | EXEC */ PAGE_EXECUTE_READWRITE, /* WRITE | EXEC */ PAGE_EXECUTE_READWRITE, /* READ | WRITE | EXEC */ PAGE_WRITECOPY, /* WRITECOPY */ PAGE_WRITECOPY, /* READ | WRITECOPY */ PAGE_WRITECOPY, /* WRITE | WRITECOPY */ PAGE_WRITECOPY, /* READ | WRITE | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* READ | EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY, /* WRITE | EXEC | WRITECOPY */ PAGE_EXECUTE_WRITECOPY /* READ | WRITE | EXEC | WRITECOPY */ }; static struct list views_list = LIST_INIT(views_list); static RTL_CRITICAL_SECTION csVirtual; static RTL_CRITICAL_SECTION_DEBUG critsect_debug = { 0, 0, &csVirtual, { &critsect_debug.ProcessLocksList, &critsect_debug.ProcessLocksList }, 0, 0, { (DWORD_PTR)(__FILE__ ": csVirtual") } }; static RTL_CRITICAL_SECTION csVirtual = { &critsect_debug, -1, 0, 0, 0, 0 }; #ifdef __i386__ /* These are always the same on an i386, and it will be faster this way */ # define page_mask 0xfff # define page_shift 12 # define page_size 0x1000 /* Note: these are Windows limits, you cannot change them. */ static void *address_space_limit = (void *)0xc0000000; /* top of the total available address space */ static void *user_space_limit = (void *)0x7fff0000; /* top of the user address space */ #else static UINT page_shift; static UINT page_size; static UINT_PTR page_mask; static void * const address_space_limit = 0; /* no limit needed on other platforms */ static void * const user_space_limit = 0; /* no limit needed on other platforms */ #endif /* __i386__ */ #define ROUND_ADDR(addr,mask) \ ((void *)((UINT_PTR)(addr) & ~(UINT_PTR)(mask))) #define ROUND_SIZE(addr,size) \ (((UINT)(size) + ((UINT_PTR)(addr) & page_mask) + page_mask) & ~page_mask) #define VIRTUAL_DEBUG_DUMP_VIEW(view) \ do { if (TRACE_ON(virtual)) VIRTUAL_DumpView(view); } while (0) #define VIRTUAL_HEAP_SIZE (4*1024*1024) static HANDLE virtual_heap; static void *preload_reserve_start; static void *preload_reserve_end; static int use_locks; static int force_exec_prot; /* whether to force PROT_EXEC on all PROT_READ mmaps */ /*********************************************************************** * VIRTUAL_GetProtStr */ static const char *VIRTUAL_GetProtStr( BYTE prot ) { static char buffer[6]; buffer[0] = (prot & VPROT_COMMITTED) ? 'c' : '-'; buffer[1] = (prot & VPROT_GUARD) ? 'g' : '-'; buffer[2] = (prot & VPROT_READ) ? 'r' : '-'; buffer[3] = (prot & VPROT_WRITECOPY) ? 'W' : ((prot & VPROT_WRITE) ? 'w' : '-'); buffer[4] = (prot & VPROT_EXEC) ? 'x' : '-'; buffer[5] = 0; return buffer; } /*********************************************************************** * VIRTUAL_GetUnixProt * * Convert page protections to protection for mmap/mprotect. */ static int VIRTUAL_GetUnixProt( BYTE vprot ) { int prot = 0; if ((vprot & VPROT_COMMITTED) && !(vprot & VPROT_GUARD)) { if (vprot & VPROT_READ) prot |= PROT_READ; if (vprot & VPROT_WRITE) prot |= PROT_WRITE; if (vprot & VPROT_WRITECOPY) prot |= PROT_WRITE; if (vprot & VPROT_EXEC) prot |= PROT_EXEC; } if (!prot) prot = PROT_NONE; return prot; } /*********************************************************************** * VIRTUAL_DumpView */ static void VIRTUAL_DumpView( FILE_VIEW *view ) { UINT i, count; char *addr = view->base; BYTE prot = view->prot[0]; TRACE( "View: %p - %p", addr, addr + view->size - 1 ); if (view->protect & VPROT_SYSTEM) TRACE( " (system)\n" ); else if (view->protect & VPROT_VALLOC) TRACE( " (valloc)\n" ); else if (view->mapping) TRACE( " %p\n", view->mapping ); else TRACE( " (anonymous)\n"); for (count = i = 1; i < view->size >> page_shift; i++, count++) { if (view->prot[i] == prot) continue; TRACE( " %p - %p %s\n", addr, addr + (count << page_shift) - 1, VIRTUAL_GetProtStr(prot) ); addr += (count << page_shift); prot = view->prot[i]; count = 0; } if (count) TRACE( " %p - %p %s\n", addr, addr + (count << page_shift) - 1, VIRTUAL_GetProtStr(prot) ); } /*********************************************************************** * VIRTUAL_Dump */ #if WINE_VM_DEBUG static void VIRTUAL_Dump(void) { sigset_t sigset; struct file_view *view; TRACE( "Dump of all virtual memory views:\n" ); server_enter_uninterrupted_section( &csVirtual, &sigset ); LIST_FOR_EACH_ENTRY( view, &views_list, FILE_VIEW, entry ) { VIRTUAL_DumpView( view ); } server_leave_uninterrupted_section( &csVirtual, &sigset ); } #endif /*********************************************************************** * VIRTUAL_FindView * * Find the view containing a given address. The csVirtual section must be held by caller. * * PARAMS * addr [I] Address * * RETURNS * View: Success * NULL: Failure */ static struct file_view *VIRTUAL_FindView( const void *addr ) { struct file_view *view; LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry ) { if (view->base > addr) break; if ((const char*)view->base + view->size > (const char*)addr) return view; } return NULL; } /*********************************************************************** * get_mask */ static inline UINT_PTR get_mask( ULONG zero_bits ) { if (!zero_bits) return 0xffff; /* allocations are aligned to 64K by default */ if (zero_bits < page_shift) zero_bits = page_shift; return (1 << zero_bits) - 1; } /*********************************************************************** * find_view_range * * Find the first view overlapping at least part of the specified range. * The csVirtual section must be held by caller. */ static struct file_view *find_view_range( const void *addr, size_t size ) { struct file_view *view; LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry ) { if ((const char *)view->base >= (const char *)addr + size) break; if ((const char *)view->base + view->size > (const char *)addr) return view; } return NULL; } /*********************************************************************** * find_free_area * * Find a free area between views inside the specified range. * The csVirtual section must be held by caller. */ static void *find_free_area( void *base, void *end, size_t size, size_t mask, int top_down ) { struct list *ptr; void *start; if (top_down) { start = ROUND_ADDR( (char *)end - size, mask ); if (start >= end || start < base) return NULL; for (ptr = views_list.prev; ptr != &views_list; ptr = ptr->prev) { struct file_view *view = LIST_ENTRY( ptr, struct file_view, entry ); if ((char *)view->base + view->size <= (char *)start) break; if ((char *)view->base >= (char *)start + size) continue; start = ROUND_ADDR( (char *)view->base - size, mask ); /* stop if remaining space is not large enough */ if (!start || start >= end || start < base) return NULL; } } else { start = ROUND_ADDR( (char *)base + mask, mask ); if (start >= end || (char *)end - (char *)start < size) return NULL; for (ptr = views_list.next; ptr != &views_list; ptr = ptr->next) { struct file_view *view = LIST_ENTRY( ptr, struct file_view, entry ); if ((char *)view->base >= (char *)start + size) break; if ((char *)view->base + view->size <= (char *)start) continue; start = ROUND_ADDR( (char *)view->base + view->size + mask, mask ); /* stop if remaining space is not large enough */ if (!start || start >= end || (char *)end - (char *)start < size) return NULL; } } return start; } /*********************************************************************** * add_reserved_area * * Add a reserved area to the list maintained by libwine. * The csVirtual section must be held by caller. */ static void add_reserved_area( void *addr, size_t size ) { TRACE( "adding %p-%p\n", addr, (char *)addr + size ); if (addr < user_space_limit) { /* unmap the part of the area that is below the limit */ assert( (char *)addr + size > (char *)user_space_limit ); munmap( addr, (char *)user_space_limit - (char *)addr ); size -= (char *)user_space_limit - (char *)addr; addr = user_space_limit; } /* blow away existing mappings */ wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED ); wine_mmap_add_reserved_area( addr, size ); } /*********************************************************************** * is_beyond_limit * * Check if an address range goes beyond a given limit. */ static inline int is_beyond_limit( const void *addr, size_t size, const void *limit ) { return (limit && (addr >= limit || (const char *)addr + size > (const char *)limit)); } /*********************************************************************** * unmap_area * * Unmap an area, or simply replace it by an empty mapping if it is * in a reserved area. The csVirtual section must be held by caller. */ static inline void unmap_area( void *addr, size_t size ) { if (wine_mmap_is_in_reserved_area( addr, size )) wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED ); else if (is_beyond_limit( addr, size, user_space_limit )) add_reserved_area( addr, size ); else munmap( addr, size ); } /*********************************************************************** * delete_view * * Deletes a view. The csVirtual section must be held by caller. */ static void delete_view( struct file_view *view ) /* [in] View */ { if (!(view->protect & VPROT_SYSTEM)) unmap_area( view->base, view->size ); list_remove( &view->entry ); if (view->mapping) NtClose( view->mapping ); RtlFreeHeap( virtual_heap, 0, view ); } /*********************************************************************** * create_view * * Create a view. The csVirtual section must be held by caller. */ static NTSTATUS create_view( struct file_view **view_ret, void *base, size_t size, unsigned int vprot ) { struct file_view *view; struct list *ptr; int unix_prot = VIRTUAL_GetUnixProt( vprot ); assert( !((UINT_PTR)base & page_mask) ); assert( !(size & page_mask) ); /* Create the view structure */ if (!(view = RtlAllocateHeap( virtual_heap, 0, sizeof(*view) + (size >> page_shift) - 1 ))) { FIXME( "out of memory in virtual heap for %p-%p\n", base, (char *)base + size ); return STATUS_NO_MEMORY; } view->base = base; view->size = size; view->mapping = 0; view->protect = vprot; memset( view->prot, vprot, size >> page_shift ); /* Insert it in the linked list */ LIST_FOR_EACH( ptr, &views_list ) { struct file_view *next = LIST_ENTRY( ptr, struct file_view, entry ); if (next->base > base) break; } list_add_before( ptr, &view->entry ); /* Check for overlapping views. This can happen if the previous view * was a system view that got unmapped behind our back. In that case * we recover by simply deleting it. */ if ((ptr = list_prev( &views_list, &view->entry )) != NULL) { struct file_view *prev = LIST_ENTRY( ptr, struct file_view, entry ); if ((char *)prev->base + prev->size > (char *)base) { TRACE( "overlapping prev view %p-%p for %p-%p\n", prev->base, (char *)prev->base + prev->size, base, (char *)base + view->size ); assert( prev->protect & VPROT_SYSTEM ); delete_view( prev ); } } if ((ptr = list_next( &views_list, &view->entry )) != NULL) { struct file_view *next = LIST_ENTRY( ptr, struct file_view, entry ); if ((char *)base + view->size > (char *)next->base) { TRACE( "overlapping next view %p-%p for %p-%p\n", next->base, (char *)next->base + next->size, base, (char *)base + view->size ); assert( next->protect & VPROT_SYSTEM ); delete_view( next ); } } *view_ret = view; VIRTUAL_DEBUG_DUMP_VIEW( view ); if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 ); mprotect( base, size, unix_prot | PROT_EXEC ); } return STATUS_SUCCESS; } /*********************************************************************** * VIRTUAL_GetWin32Prot * * Convert page protections to Win32 flags. */ static DWORD VIRTUAL_GetWin32Prot( BYTE vprot ) { DWORD ret = VIRTUAL_Win32Flags[vprot & 0x0f]; if (vprot & VPROT_NOCACHE) ret |= PAGE_NOCACHE; if (vprot & VPROT_GUARD) ret |= PAGE_GUARD; return ret; } /*********************************************************************** * VIRTUAL_GetProt * * Build page protections from Win32 flags. * * PARAMS * protect [I] Win32 protection flags * * RETURNS * Value of page protection flags */ static BYTE VIRTUAL_GetProt( DWORD protect ) { BYTE vprot; switch(protect & 0xff) { case PAGE_READONLY: vprot = VPROT_READ; break; case PAGE_READWRITE: vprot = VPROT_READ | VPROT_WRITE; break; case PAGE_WRITECOPY: /* MSDN CreateFileMapping() states that if PAGE_WRITECOPY is given, * that the hFile must have been opened with GENERIC_READ and * GENERIC_WRITE access. This is WRONG as tests show that you * only need GENERIC_READ access (at least for Win9x, * FIXME: what about NT?). Thus, we don't put VPROT_WRITE in * PAGE_WRITECOPY and PAGE_EXECUTE_WRITECOPY. */ vprot = VPROT_READ | VPROT_WRITECOPY; break; case PAGE_EXECUTE: vprot = VPROT_EXEC; break; case PAGE_EXECUTE_READ: vprot = VPROT_EXEC | VPROT_READ; break; case PAGE_EXECUTE_READWRITE: vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITE; break; case PAGE_EXECUTE_WRITECOPY: /* See comment for PAGE_WRITECOPY above */ vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITECOPY; break; case PAGE_NOACCESS: default: vprot = 0; break; } if (protect & PAGE_GUARD) vprot |= VPROT_GUARD; if (protect & PAGE_NOCACHE) vprot |= VPROT_NOCACHE; return vprot; } /*********************************************************************** * VIRTUAL_SetProt * * Change the protection of a range of pages. * * RETURNS * TRUE: Success * FALSE: Failure */ static BOOL VIRTUAL_SetProt( FILE_VIEW *view, /* [in] Pointer to view */ void *base, /* [in] Starting address */ size_t size, /* [in] Size in bytes */ BYTE vprot ) /* [in] Protections to use */ { int unix_prot = VIRTUAL_GetUnixProt(vprot); TRACE("%p-%p %s\n", base, (char *)base + size - 1, VIRTUAL_GetProtStr( vprot ) ); /* if setting stack guard pages, store the permissions first, as the guard may be * triggered at any point after mprotect and change the permissions again */ if ((vprot & VPROT_GUARD) && ((char *)base >= (char *)NtCurrentTeb()->DeallocationStack) && ((char *)base < (char *)NtCurrentTeb()->Tib.StackBase)) { memset( view->prot + (((char *)base - (char *)view->base) >> page_shift), vprot, size >> page_shift ); mprotect( base, size, unix_prot ); VIRTUAL_DEBUG_DUMP_VIEW( view ); return TRUE; } if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 ); if (!mprotect( base, size, unix_prot | PROT_EXEC )) goto done; /* exec + write may legitimately fail, in that case fall back to write only */ if (!(unix_prot & PROT_WRITE)) return FALSE; } if (mprotect( base, size, unix_prot )) return FALSE; /* FIXME: last error */ done: memset( view->prot + (((char *)base - (char *)view->base) >> page_shift), vprot, size >> page_shift ); VIRTUAL_DEBUG_DUMP_VIEW( view ); return TRUE; } /*********************************************************************** * unmap_extra_space * * Release the extra memory while keeping the range starting on the granularity boundary. */ static inline void *unmap_extra_space( void *ptr, size_t total_size, size_t wanted_size, size_t mask ) { if ((ULONG_PTR)ptr & mask) { size_t extra = mask + 1 - ((ULONG_PTR)ptr & mask); munmap( ptr, extra ); ptr = (char *)ptr + extra; total_size -= extra; } if (total_size > wanted_size) munmap( (char *)ptr + wanted_size, total_size - wanted_size ); return ptr; } struct alloc_area { size_t size; size_t mask; int top_down; void *result; }; /*********************************************************************** * alloc_reserved_area_callback * * Try to map some space inside a reserved area. Callback for wine_mmap_enum_reserved_areas. */ static int alloc_reserved_area_callback( void *start, size_t size, void *arg ) { static void * const address_space_start = (void *)0x110000; struct alloc_area *alloc = arg; void *end = (char *)start + size; if (start < address_space_start) start = address_space_start; if (user_space_limit && end > user_space_limit) end = user_space_limit; if (start >= end) return 0; /* make sure we don't touch the preloader reserved range */ if (preload_reserve_end >= start) { if (preload_reserve_end >= end) { if (preload_reserve_start <= start) return 0; /* no space in that area */ if (preload_reserve_start < end) end = preload_reserve_start; } else if (preload_reserve_start <= start) start = preload_reserve_end; else { /* range is split in two by the preloader reservation, try first part */ if ((alloc->result = find_free_area( start, preload_reserve_start, alloc->size, alloc->mask, alloc->top_down ))) return 1; /* then fall through to try second part */ start = preload_reserve_end; } } if ((alloc->result = find_free_area( start, end, alloc->size, alloc->mask, alloc->top_down ))) return 1; return 0; } /*********************************************************************** * map_view * * Create a view and mmap the corresponding memory area. * The csVirtual section must be held by caller. */ static NTSTATUS map_view( struct file_view **view_ret, void *base, size_t size, size_t mask, int top_down, unsigned int vprot ) { void *ptr; NTSTATUS status; if (base) { if (is_beyond_limit( base, size, address_space_limit )) return STATUS_WORKING_SET_LIMIT_RANGE; switch (wine_mmap_is_in_reserved_area( base, size )) { case -1: /* partially in a reserved area */ return STATUS_CONFLICTING_ADDRESSES; case 0: /* not in a reserved area, do a normal allocation */ if ((ptr = wine_anon_mmap( base, size, VIRTUAL_GetUnixProt(vprot), 0 )) == (void *)-1) { if (errno == ENOMEM) return STATUS_NO_MEMORY; return STATUS_INVALID_PARAMETER; } if (ptr != base) { /* We couldn't get the address we wanted */ if (is_beyond_limit( ptr, size, user_space_limit )) add_reserved_area( ptr, size ); else munmap( ptr, size ); return STATUS_CONFLICTING_ADDRESSES; } break; default: case 1: /* in a reserved area, make sure the address is available */ if (find_view_range( base, size )) return STATUS_CONFLICTING_ADDRESSES; /* replace the reserved area by our mapping */ if ((ptr = wine_anon_mmap( base, size, VIRTUAL_GetUnixProt(vprot), MAP_FIXED )) != base) return STATUS_INVALID_PARAMETER; break; } } else { size_t view_size = size + mask + 1; struct alloc_area alloc; alloc.size = size; alloc.mask = mask; alloc.top_down = top_down; if (wine_mmap_enum_reserved_areas( alloc_reserved_area_callback, &alloc, top_down )) { ptr = alloc.result; TRACE( "got mem in reserved area %p-%p\n", ptr, (char *)ptr + size ); if (wine_anon_mmap( ptr, size, VIRTUAL_GetUnixProt(vprot), MAP_FIXED ) != ptr) return STATUS_INVALID_PARAMETER; goto done; } for (;;) { if ((ptr = wine_anon_mmap( NULL, view_size, VIRTUAL_GetUnixProt(vprot), 0 )) == (void *)-1) { if (errno == ENOMEM) return STATUS_NO_MEMORY; return STATUS_INVALID_PARAMETER; } TRACE( "got mem with anon mmap %p-%p\n", ptr, (char *)ptr + size ); /* if we got something beyond the user limit, unmap it and retry */ if (is_beyond_limit( ptr, view_size, user_space_limit )) add_reserved_area( ptr, view_size ); else break; } ptr = unmap_extra_space( ptr, view_size, size, mask ); } done: status = create_view( view_ret, ptr, size, vprot ); if (status != STATUS_SUCCESS) unmap_area( ptr, size ); return status; } /*********************************************************************** * unaligned_mmap * * Linux kernels before 2.4.x can support non page-aligned offsets, as * long as the offset is aligned to the filesystem block size. This is * a big performance gain so we want to take advantage of it. * * However, when we use 64-bit file support this doesn't work because * glibc rejects unaligned offsets. Also glibc 2.1.3 mmap64 is broken * in that it rounds unaligned offsets down to a page boundary. For * these reasons we do a direct system call here. */ static void *unaligned_mmap( void *addr, size_t length, unsigned int prot, unsigned int flags, int fd, off_t offset ) { #if defined(linux) && defined(__i386__) && defined(__GNUC__) if (!(offset >> 32) && (offset & page_mask)) { int ret; struct { void *addr; unsigned int length; unsigned int prot; unsigned int flags; unsigned int fd; unsigned int offset; } args; args.addr = addr; args.length = length; args.prot = prot; args.flags = flags; args.fd = fd; args.offset = offset; __asm__ __volatile__("push %%ebx\n\t" "movl %2,%%ebx\n\t" "int $0x80\n\t" "popl %%ebx" : "=a" (ret) : "0" (90), /* SYS_mmap */ "q" (&args) : "memory" ); if (ret < 0 && ret > -4096) { errno = -ret; ret = -1; } return (void *)ret; } #endif return mmap( addr, length, prot, flags, fd, offset ); } /*********************************************************************** * map_file_into_view * * Wrapper for mmap() to map a file into a view, falling back to read if mmap fails. * The csVirtual section must be held by caller. */ static NTSTATUS map_file_into_view( struct file_view *view, int fd, size_t start, size_t size, off_t offset, unsigned int vprot, BOOL removable ) { void *ptr; int prot = VIRTUAL_GetUnixProt( vprot ); BOOL shared_write = (vprot & VPROT_WRITE) != 0; assert( start < view->size ); assert( start + size <= view->size ); /* only try mmap if media is not removable (or if we require write access) */ if (!removable || shared_write) { int flags = MAP_FIXED | (shared_write ? MAP_SHARED : MAP_PRIVATE); if (unaligned_mmap( (char *)view->base + start, size, prot, flags, fd, offset ) != (void *)-1) goto done; /* mmap() failed; if this is because the file offset is not */ /* page-aligned (EINVAL), or because the underlying filesystem */ /* does not support mmap() (ENOEXEC,ENODEV), we do it by hand. */ if ((errno != ENOEXEC) && (errno != EINVAL) && (errno != ENODEV)) return FILE_GetNtStatus(); if (shared_write) /* we cannot fake shared write mappings */ { if (errno == EINVAL) return STATUS_INVALID_PARAMETER; ERR( "shared writable mmap not supported, broken filesystem?\n" ); return STATUS_NOT_SUPPORTED; } } /* Reserve the memory with an anonymous mmap */ ptr = wine_anon_mmap( (char *)view->base + start, size, PROT_READ | PROT_WRITE, MAP_FIXED ); if (ptr == (void *)-1) return FILE_GetNtStatus(); /* Now read in the file */ pread( fd, ptr, size, offset ); if (prot != (PROT_READ|PROT_WRITE)) mprotect( ptr, size, prot ); /* Set the right protection */ done: memset( view->prot + (start >> page_shift), vprot, ROUND_SIZE(start,size) >> page_shift ); return STATUS_SUCCESS; } /*********************************************************************** * decommit_view * * Decommit some pages of a given view. * The csVirtual section must be held by caller. */ static NTSTATUS decommit_pages( struct file_view *view, size_t start, size_t size ) { if (wine_anon_mmap( (char *)view->base + start, size, PROT_NONE, MAP_FIXED ) != (void *)-1) { BYTE *p = view->prot + (start >> page_shift); size >>= page_shift; while (size--) *p++ &= ~VPROT_COMMITTED; return STATUS_SUCCESS; } return FILE_GetNtStatus(); } /*********************************************************************** * map_image * * Map an executable (PE format) image into memory. */ static NTSTATUS map_image( HANDLE hmapping, int fd, char *base, SIZE_T total_size, SIZE_T mask, SIZE_T header_size, int shared_fd, HANDLE dup_mapping, PVOID *addr_ptr ) { IMAGE_DOS_HEADER *dos; IMAGE_NT_HEADERS *nt; IMAGE_SECTION_HEADER *sec; IMAGE_DATA_DIRECTORY *imports; NTSTATUS status = STATUS_CONFLICTING_ADDRESSES; int i; off_t pos; sigset_t sigset; struct stat st; struct file_view *view = NULL; char *ptr, *header_end; int delta = 0; /* zero-map the whole range */ server_enter_uninterrupted_section( &csVirtual, &sigset ); if (base >= (char *)0x110000) /* make sure the DOS area remains free */ status = map_view( &view, base, total_size, mask, FALSE, VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY | VPROT_IMAGE ); if (status == STATUS_CONFLICTING_ADDRESSES) status = map_view( &view, NULL, total_size, mask, FALSE, VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY | VPROT_IMAGE ); if (status != STATUS_SUCCESS) goto error; ptr = view->base; TRACE_(module)( "mapped PE file at %p-%p\n", ptr, ptr + total_size ); /* map the header */ if (fstat( fd, &st ) == -1) { status = FILE_GetNtStatus(); goto error; } status = STATUS_INVALID_IMAGE_FORMAT; /* generic error */ if (!st.st_size) goto error; header_size = min( header_size, st.st_size ); if (map_file_into_view( view, fd, 0, header_size, 0, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, !dup_mapping ) != STATUS_SUCCESS) goto error; dos = (IMAGE_DOS_HEADER *)ptr; nt = (IMAGE_NT_HEADERS *)(ptr + dos->e_lfanew); header_end = ptr + ROUND_SIZE( 0, header_size ); memset( ptr + header_size, 0, header_end - (ptr + header_size) ); if ((char *)(nt + 1) > header_end) goto error; sec = (IMAGE_SECTION_HEADER*)((char*)&nt->OptionalHeader+nt->FileHeader.SizeOfOptionalHeader); if ((char *)(sec + nt->FileHeader.NumberOfSections) > header_end) goto error; imports = nt->OptionalHeader.DataDirectory + IMAGE_DIRECTORY_ENTRY_IMPORT; if (!imports->Size || !imports->VirtualAddress) imports = NULL; /* check the architecture */ if (nt->FileHeader.Machine != IMAGE_FILE_MACHINE_I386) { MESSAGE("Trying to load PE image for unsupported architecture ("); switch (nt->FileHeader.Machine) { case IMAGE_FILE_MACHINE_UNKNOWN: MESSAGE("Unknown"); break; case IMAGE_FILE_MACHINE_I860: MESSAGE("I860"); break; case IMAGE_FILE_MACHINE_R3000: MESSAGE("R3000"); break; case IMAGE_FILE_MACHINE_R4000: MESSAGE("R4000"); break; case IMAGE_FILE_MACHINE_R10000: MESSAGE("R10000"); break; case IMAGE_FILE_MACHINE_ALPHA: MESSAGE("Alpha"); break; case IMAGE_FILE_MACHINE_POWERPC: MESSAGE("PowerPC"); break; case IMAGE_FILE_MACHINE_IA64: MESSAGE("IA-64"); break; case IMAGE_FILE_MACHINE_ALPHA64: MESSAGE("Alpha-64"); break; case IMAGE_FILE_MACHINE_AMD64: MESSAGE("AMD-64"); break; case IMAGE_FILE_MACHINE_ARM: MESSAGE("ARM"); break; default: MESSAGE("Unknown-%04x", nt->FileHeader.Machine); break; } MESSAGE(")\n"); goto error; } /* check for non page-aligned binary */ if (nt->OptionalHeader.SectionAlignment <= page_mask) { /* unaligned sections, this happens for native subsystem binaries */ /* in that case Windows simply maps in the whole file */ if (map_file_into_view( view, fd, 0, total_size, 0, VPROT_COMMITTED | VPROT_READ, !dup_mapping ) != STATUS_SUCCESS) goto error; /* check that all sections are loaded at the right offset */ if (nt->OptionalHeader.FileAlignment != nt->OptionalHeader.SectionAlignment) goto error; for (i = 0; i < nt->FileHeader.NumberOfSections; i++) { if (sec[i].VirtualAddress != sec[i].PointerToRawData) goto error; /* Windows refuses to load in that case too */ } /* set the image protections */ VIRTUAL_SetProt( view, ptr, total_size, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY | VPROT_EXEC ); /* no relocations are performed on non page-aligned binaries */ goto done; } /* map all the sections */ for (i = pos = 0; i < nt->FileHeader.NumberOfSections; i++, sec++) { static const SIZE_T sector_align = 0x1ff; SIZE_T map_size, file_start, file_size, end; if (!sec->Misc.VirtualSize) map_size = ROUND_SIZE( 0, sec->SizeOfRawData ); else map_size = ROUND_SIZE( 0, sec->Misc.VirtualSize ); /* file positions are rounded to sector boundaries regardless of OptionalHeader.FileAlignment */ file_start = sec->PointerToRawData & ~sector_align; file_size = (sec->SizeOfRawData + (sec->PointerToRawData & sector_align) + sector_align) & ~sector_align; if (file_size > map_size) file_size = map_size; /* a few sanity checks */ end = sec->VirtualAddress + ROUND_SIZE( sec->VirtualAddress, map_size ); if (sec->VirtualAddress > total_size || end > total_size || end < sec->VirtualAddress) { WARN_(module)( "Section %.8s too large (%x+%lx/%lx)\n", sec->Name, sec->VirtualAddress, map_size, total_size ); goto error; } if ((sec->Characteristics & IMAGE_SCN_MEM_SHARED) && (sec->Characteristics & IMAGE_SCN_MEM_WRITE)) { TRACE_(module)( "mapping shared section %.8s at %p off %x (%x) size %lx (%lx) flags %x\n", sec->Name, ptr + sec->VirtualAddress, sec->PointerToRawData, (int)pos, file_size, map_size, sec->Characteristics ); if (map_file_into_view( view, shared_fd, sec->VirtualAddress, map_size, pos, VPROT_COMMITTED | VPROT_READ | VPROT_WRITE, FALSE ) != STATUS_SUCCESS) { ERR_(module)( "Could not map shared section %.8s\n", sec->Name ); goto error; } /* check if the import directory falls inside this section */ if (imports && imports->VirtualAddress >= sec->VirtualAddress && imports->VirtualAddress < sec->VirtualAddress + map_size) { UINT_PTR base = imports->VirtualAddress & ~page_mask; UINT_PTR end = base + ROUND_SIZE( imports->VirtualAddress, imports->Size ); if (end > sec->VirtualAddress + map_size) end = sec->VirtualAddress + map_size; if (end > base) map_file_into_view( view, shared_fd, base, end - base, pos + (base - sec->VirtualAddress), VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, FALSE ); } pos += map_size; continue; } TRACE_(module)( "mapping section %.8s at %p off %x size %x virt %x flags %x\n", sec->Name, ptr + sec->VirtualAddress, sec->PointerToRawData, sec->SizeOfRawData, sec->Misc.VirtualSize, sec->Characteristics ); if (!sec->PointerToRawData || !file_size) continue; /* Note: if the section is not aligned properly map_file_into_view will magically * fall back to read(), so we don't need to check anything here. */ end = file_start + file_size; if (sec->PointerToRawData >= st.st_size || end > ((st.st_size + sector_align) & ~sector_align) || end < file_start || map_file_into_view( view, fd, sec->VirtualAddress, file_size, file_start, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY, !dup_mapping ) != STATUS_SUCCESS) { ERR_(module)( "Could not map section %.8s, file probably truncated\n", sec->Name ); goto error; } if (file_size & page_mask) { end = ROUND_SIZE( 0, file_size ); if (end > map_size) end = map_size; TRACE_(module)("clearing %p - %p\n", ptr + sec->VirtualAddress + file_size, ptr + sec->VirtualAddress + end ); memset( ptr + sec->VirtualAddress + file_size, 0, end - file_size ); } } /* perform base relocation, if necessary */ if (ptr != base && ((nt->FileHeader.Characteristics & IMAGE_FILE_DLL) || !NtCurrentTeb()->Peb->ImageBaseAddress) ) { IMAGE_BASE_RELOCATION *rel, *end; const IMAGE_DATA_DIRECTORY *relocs; if (nt->FileHeader.Characteristics & IMAGE_FILE_RELOCS_STRIPPED) { WARN_(module)( "Need to relocate module from %p to %p, but there are no relocation records\n", base, ptr ); status = STATUS_CONFLICTING_ADDRESSES; goto error; } TRACE_(module)( "relocating from %p-%p to %p-%p\n", base, base + total_size, ptr, ptr + total_size ); relocs = &nt->OptionalHeader.DataDirectory[IMAGE_DIRECTORY_ENTRY_BASERELOC]; rel = (IMAGE_BASE_RELOCATION *)(ptr + relocs->VirtualAddress); end = (IMAGE_BASE_RELOCATION *)(ptr + relocs->VirtualAddress + relocs->Size); delta = ptr - base; while (rel < end - 1 && rel->SizeOfBlock) { if (rel->VirtualAddress >= total_size) { WARN_(module)( "invalid address %p in relocation %p\n", ptr + rel->VirtualAddress, rel ); status = STATUS_ACCESS_VIOLATION; goto error; } rel = LdrProcessRelocationBlock( ptr + rel->VirtualAddress, (rel->SizeOfBlock - sizeof(*rel)) / sizeof(USHORT), (USHORT *)(rel + 1), delta ); if (!rel) goto error; } } /* set the image protections */ VIRTUAL_SetProt( view, ptr, ROUND_SIZE( 0, header_size ), VPROT_COMMITTED | VPROT_READ ); sec = (IMAGE_SECTION_HEADER*)((char *)&nt->OptionalHeader+nt->FileHeader.SizeOfOptionalHeader); for (i = 0; i < nt->FileHeader.NumberOfSections; i++, sec++) { SIZE_T size; BYTE vprot = VPROT_COMMITTED; if (sec->Misc.VirtualSize) size = ROUND_SIZE( sec->VirtualAddress, sec->Misc.VirtualSize ); else size = ROUND_SIZE( sec->VirtualAddress, sec->SizeOfRawData ); if (sec->Characteristics & IMAGE_SCN_MEM_READ) vprot |= VPROT_READ; if (sec->Characteristics & IMAGE_SCN_MEM_WRITE) vprot |= VPROT_READ|VPROT_WRITECOPY; if (sec->Characteristics & IMAGE_SCN_MEM_EXECUTE) vprot |= VPROT_EXEC; /* Dumb game crack lets the AOEP point into a data section. Adjust. */ if ((nt->OptionalHeader.AddressOfEntryPoint >= sec->VirtualAddress) && (nt->OptionalHeader.AddressOfEntryPoint < sec->VirtualAddress + size)) vprot |= VPROT_EXEC; VIRTUAL_SetProt( view, ptr + sec->VirtualAddress, size, vprot ); } done: view->mapping = dup_mapping; server_leave_uninterrupted_section( &csVirtual, &sigset ); *addr_ptr = ptr; #ifdef VALGRIND_LOAD_PDB_DEBUGINFO VALGRIND_LOAD_PDB_DEBUGINFO(fd, ptr, total_size, delta); #endif return STATUS_SUCCESS; error: if (view) delete_view( view ); server_leave_uninterrupted_section( &csVirtual, &sigset ); if (dup_mapping) NtClose( dup_mapping ); return status; } /* callback for wine_mmap_enum_reserved_areas to allocate space for the virtual heap */ static int alloc_virtual_heap( void *base, size_t size, void *arg ) { void **heap_base = arg; if (address_space_limit) address_space_limit = max( (char *)address_space_limit, (char *)base + size ); if (size < VIRTUAL_HEAP_SIZE) return 0; *heap_base = wine_anon_mmap( (char *)base + size - VIRTUAL_HEAP_SIZE, VIRTUAL_HEAP_SIZE, PROT_READ|PROT_WRITE, MAP_FIXED ); return (*heap_base != (void *)-1); } /*********************************************************************** * virtual_init */ void virtual_init(void) { const char *preload; void *heap_base; struct file_view *heap_view; #ifndef page_mask page_size = getpagesize(); page_mask = page_size - 1; /* Make sure we have a power of 2 */ assert( !(page_size & page_mask) ); page_shift = 0; while ((1 << page_shift) != page_size) page_shift++; #endif /* page_mask */ if ((preload = getenv("WINEPRELOADRESERVE"))) { unsigned long start, end; if (sscanf( preload, "%lx-%lx", &start, &end ) == 2) { preload_reserve_start = (void *)start; preload_reserve_end = (void *)end; } } /* try to find space in a reserved area for the virtual heap */ if (!wine_mmap_enum_reserved_areas( alloc_virtual_heap, &heap_base, 1 )) heap_base = wine_anon_mmap( NULL, VIRTUAL_HEAP_SIZE, PROT_READ|PROT_WRITE, 0 ); assert( heap_base != (void *)-1 ); virtual_heap = RtlCreateHeap( HEAP_NO_SERIALIZE, heap_base, VIRTUAL_HEAP_SIZE, VIRTUAL_HEAP_SIZE, NULL, NULL ); create_view( &heap_view, heap_base, VIRTUAL_HEAP_SIZE, VPROT_COMMITTED | VPROT_READ | VPROT_WRITE ); } /*********************************************************************** * virtual_init_threading */ void virtual_init_threading(void) { use_locks = 1; } /*********************************************************************** * virtual_get_system_info */ void virtual_get_system_info( SYSTEM_BASIC_INFORMATION *info ) { info->dwUnknown1 = 0; info->uKeMaximumIncrement = 0; /* FIXME */ info->uPageSize = page_size; info->uMmLowestPhysicalPage = 1; info->uMmHighestPhysicalPage = 0x7fffffff / page_size; info->uMmNumberOfPhysicalPages = info->uMmHighestPhysicalPage - info->uMmLowestPhysicalPage; info->uAllocationGranularity = get_mask(0) + 1; info->pLowestUserAddress = (void *)0x10000; info->pMmHighestUserAddress = (char *)user_space_limit - 1; info->uKeActiveProcessors = NtCurrentTeb()->Peb->NumberOfProcessors; info->bKeNumberProcessors = info->uKeActiveProcessors; } /*********************************************************************** * virtual_alloc_thread_stack */ NTSTATUS virtual_alloc_thread_stack( void *base, SIZE_T size ) { FILE_VIEW *view; NTSTATUS status; sigset_t sigset; server_enter_uninterrupted_section( &csVirtual, &sigset ); if (base) /* already allocated, create a system view */ { size = ROUND_SIZE( base, size ); base = ROUND_ADDR( base, page_mask ); if ((status = create_view( &view, base, size, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED | VPROT_VALLOC | VPROT_SYSTEM )) != STATUS_SUCCESS) goto done; } else { size = (size + 0xffff) & ~0xffff; /* round to 64K boundary */ if ((status = map_view( &view, NULL, size, 0xffff, 0, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED | VPROT_VALLOC )) != STATUS_SUCCESS) goto done; #ifdef VALGRIND_STACK_REGISTER /* no need to de-register the stack as it's the one of the main thread */ VALGRIND_STACK_REGISTER( view->base, (char *)view->base + view->size ); #endif } /* setup no access guard page */ VIRTUAL_SetProt( view, view->base, page_size, VPROT_COMMITTED ); VIRTUAL_SetProt( view, (char *)view->base + page_size, page_size, VPROT_READ | VPROT_WRITE | VPROT_COMMITTED | VPROT_GUARD ); /* note: limit is lower than base since the stack grows down */ NtCurrentTeb()->DeallocationStack = view->base; NtCurrentTeb()->Tib.StackBase = (char *)view->base + view->size; NtCurrentTeb()->Tib.StackLimit = (char *)view->base + 2 * page_size; done: server_leave_uninterrupted_section( &csVirtual, &sigset ); return status; } /*********************************************************************** * virtual_clear_thread_stack * * Clear the stack contents before calling the main entry point, some broken apps need that. */ void virtual_clear_thread_stack(void) { void *stack = NtCurrentTeb()->Tib.StackLimit; size_t size = (char *)NtCurrentTeb()->Tib.StackBase - (char *)NtCurrentTeb()->Tib.StackLimit; wine_anon_mmap( stack, size, PROT_READ | PROT_WRITE, MAP_FIXED ); if (force_exec_prot) mprotect( stack, size, PROT_READ | PROT_WRITE | PROT_EXEC ); } /*********************************************************************** * VIRTUAL_HandleFault */ NTSTATUS VIRTUAL_HandleFault( LPCVOID addr ) { FILE_VIEW *view; NTSTATUS ret = STATUS_ACCESS_VIOLATION; sigset_t sigset; server_enter_uninterrupted_section( &csVirtual, &sigset ); if ((view = VIRTUAL_FindView( addr ))) { void *page = ROUND_ADDR( addr, page_mask ); BYTE vprot = view->prot[((const char *)page - (const char *)view->base) >> page_shift]; if (vprot & VPROT_GUARD) { VIRTUAL_SetProt( view, page, page_size, vprot & ~VPROT_GUARD ); ret = STATUS_GUARD_PAGE_VIOLATION; } } server_leave_uninterrupted_section( &csVirtual, &sigset ); return ret; } /*********************************************************************** * virtual_handle_stack_fault * * Handle an access fault inside the current thread stack. * Called from inside a signal handler. */ BOOL virtual_handle_stack_fault( void *addr ) { FILE_VIEW *view; BOOL ret = FALSE; RtlEnterCriticalSection( &csVirtual ); /* no need for signal masking inside signal handler */ if ((view = VIRTUAL_FindView( addr ))) { void *page = ROUND_ADDR( addr, page_mask ); BYTE vprot = view->prot[((const char *)page - (const char *)view->base) >> page_shift]; if (vprot & VPROT_GUARD) { VIRTUAL_SetProt( view, page, page_size, vprot & ~VPROT_GUARD ); if ((char *)page + page_size == NtCurrentTeb()->Tib.StackLimit) NtCurrentTeb()->Tib.StackLimit = page; ret = TRUE; } } RtlLeaveCriticalSection( &csVirtual ); return ret; } /*********************************************************************** * VIRTUAL_SetForceExec * * Whether to force exec prot on all views. */ void VIRTUAL_SetForceExec( BOOL enable ) { struct file_view *view; sigset_t sigset; server_enter_uninterrupted_section( &csVirtual, &sigset ); if (!force_exec_prot != !enable) /* change all existing views */ { force_exec_prot = enable; LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry ) { UINT i, count; int unix_prot; char *addr = view->base; BYTE prot = view->prot[0]; for (count = i = 1; i < view->size >> page_shift; i++, count++) { if (view->prot[i] == prot) continue; unix_prot = VIRTUAL_GetUnixProt( prot ); if ((unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "%s exec prot for %p-%p\n", force_exec_prot ? "enabling" : "disabling", addr, addr + (count << page_shift) - 1 ); mprotect( addr, count << page_shift, unix_prot | (force_exec_prot ? PROT_EXEC : 0) ); } addr += (count << page_shift); prot = view->prot[i]; count = 0; } if (count) { unix_prot = VIRTUAL_GetUnixProt( prot ); if ((unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC)) { TRACE( "%s exec prot for %p-%p\n", force_exec_prot ? "enabling" : "disabling", addr, addr + (count << page_shift) - 1 ); mprotect( addr, count << page_shift, unix_prot | (force_exec_prot ? PROT_EXEC : 0) ); } } } } server_leave_uninterrupted_section( &csVirtual, &sigset ); } /*********************************************************************** * VIRTUAL_UseLargeAddressSpace * * Increase the address space size for apps that support it. */ void VIRTUAL_UseLargeAddressSpace(void) { /* no large address space on win9x */ if (NtCurrentTeb()->Peb->OSPlatformId != VER_PLATFORM_WIN32_NT) return; user_space_limit = address_space_limit; } /*********************************************************************** * NtAllocateVirtualMemory (NTDLL.@) * ZwAllocateVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtAllocateVirtualMemory( HANDLE process, PVOID *ret, ULONG zero_bits, SIZE_T *size_ptr, ULONG type, ULONG protect ) { void *base; unsigned int vprot; SIZE_T size = *size_ptr; SIZE_T mask = get_mask( zero_bits ); NTSTATUS status = STATUS_SUCCESS; struct file_view *view; sigset_t sigset; TRACE("%p %p %08lx %x %08x\n", process, *ret, size, type, protect ); if (!size) return STATUS_INVALID_PARAMETER; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_alloc.type = APC_VIRTUAL_ALLOC; call.virtual_alloc.addr = *ret; call.virtual_alloc.size = *size_ptr; call.virtual_alloc.zero_bits = zero_bits; call.virtual_alloc.op_type = type; call.virtual_alloc.prot = protect; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_alloc.status == STATUS_SUCCESS) { *ret = result.virtual_alloc.addr; *size_ptr = result.virtual_alloc.size; } return result.virtual_alloc.status; } /* Round parameters to a page boundary */ if (size > 0x7fc00000) return STATUS_WORKING_SET_LIMIT_RANGE; /* 2Gb - 4Mb */ if (*ret) { if (type & MEM_RESERVE) /* Round down to 64k boundary */ base = ROUND_ADDR( *ret, mask ); else base = ROUND_ADDR( *ret, page_mask ); size = (((UINT_PTR)*ret + size + page_mask) & ~page_mask) - (UINT_PTR)base; /* disallow low 64k, wrap-around and kernel space */ if (((char *)base < (char *)0x10000) || ((char *)base + size < (char *)base) || is_beyond_limit( base, size, address_space_limit )) return STATUS_INVALID_PARAMETER; } else { base = NULL; size = (size + page_mask) & ~page_mask; } /* Compute the alloc type flags */ if (!(type & MEM_SYSTEM)) { if (!(type & (MEM_COMMIT | MEM_RESERVE)) || (type & ~(MEM_COMMIT | MEM_RESERVE | MEM_TOP_DOWN | MEM_WRITE_WATCH | MEM_RESET))) { WARN("called with wrong alloc type flags (%08x) !\n", type); return STATUS_INVALID_PARAMETER; } if (type & MEM_WRITE_WATCH) { FIXME("MEM_WRITE_WATCH type not supported\n"); return STATUS_NOT_SUPPORTED; } } vprot = VIRTUAL_GetProt( protect ) | VPROT_VALLOC; if (type & MEM_COMMIT) vprot |= VPROT_COMMITTED; /* Reserve the memory */ if (use_locks) server_enter_uninterrupted_section( &csVirtual, &sigset ); if (type & MEM_SYSTEM) { if (type & MEM_IMAGE) vprot |= VPROT_IMAGE; status = create_view( &view, base, size, vprot | VPROT_COMMITTED | VPROT_SYSTEM ); if (status == STATUS_SUCCESS) base = view->base; } else if ((type & MEM_RESERVE) || !base) { status = map_view( &view, base, size, mask, type & MEM_TOP_DOWN, vprot ); if (status == STATUS_SUCCESS) base = view->base; } else /* commit the pages */ { if (!(view = VIRTUAL_FindView( base )) || ((char *)base + size > (char *)view->base + view->size)) status = STATUS_NOT_MAPPED_VIEW; else if (!VIRTUAL_SetProt( view, base, size, vprot )) status = STATUS_ACCESS_DENIED; } if (use_locks) server_leave_uninterrupted_section( &csVirtual, &sigset ); if (status == STATUS_SUCCESS) { *ret = base; *size_ptr = size; } return status; } /*********************************************************************** * NtFreeVirtualMemory (NTDLL.@) * ZwFreeVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtFreeVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr, ULONG type ) { FILE_VIEW *view; char *base; sigset_t sigset; NTSTATUS status = STATUS_SUCCESS; LPVOID addr = *addr_ptr; SIZE_T size = *size_ptr; TRACE("%p %p %08lx %x\n", process, addr, size, type ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_free.type = APC_VIRTUAL_FREE; call.virtual_free.addr = addr; call.virtual_free.size = size; call.virtual_free.op_type = type; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_free.status == STATUS_SUCCESS) { *addr_ptr = result.virtual_free.addr; *size_ptr = result.virtual_free.size; } return result.virtual_free.status; } /* Fix the parameters */ size = ROUND_SIZE( addr, size ); base = ROUND_ADDR( addr, page_mask ); /* avoid freeing the DOS area when a broken app passes a NULL pointer */ if (!base && !(type & MEM_SYSTEM)) return STATUS_INVALID_PARAMETER; server_enter_uninterrupted_section( &csVirtual, &sigset ); if (!(view = VIRTUAL_FindView( base )) || (base + size > (char *)view->base + view->size) || !(view->protect & VPROT_VALLOC)) { status = STATUS_INVALID_PARAMETER; } else if (type & MEM_SYSTEM) { /* return the values that the caller should use to unmap the area */ *addr_ptr = view->base; if (!wine_mmap_is_in_reserved_area( view->base, view->size )) *size_ptr = view->size; else *size_ptr = 0; /* make sure we don't munmap anything from a reserved area */ view->protect |= VPROT_SYSTEM; delete_view( view ); } else if (type == MEM_RELEASE) { /* Free the pages */ if (size || (base != view->base)) status = STATUS_INVALID_PARAMETER; else { delete_view( view ); *addr_ptr = base; *size_ptr = size; } } else if (type == MEM_DECOMMIT) { status = decommit_pages( view, base - (char *)view->base, size ); if (status == STATUS_SUCCESS) { *addr_ptr = base; *size_ptr = size; } } else { WARN("called with wrong free type flags (%08x) !\n", type); status = STATUS_INVALID_PARAMETER; } server_leave_uninterrupted_section( &csVirtual, &sigset ); return status; } /*********************************************************************** * NtProtectVirtualMemory (NTDLL.@) * ZwProtectVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtProtectVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr, ULONG new_prot, ULONG *old_prot ) { FILE_VIEW *view; sigset_t sigset; NTSTATUS status = STATUS_SUCCESS; char *base; UINT i; BYTE vprot, *p; ULONG prot; SIZE_T size = *size_ptr; LPVOID addr = *addr_ptr; TRACE("%p %p %08lx %08x\n", process, addr, size, new_prot ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_protect.type = APC_VIRTUAL_PROTECT; call.virtual_protect.addr = addr; call.virtual_protect.size = size; call.virtual_protect.prot = new_prot; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_protect.status == STATUS_SUCCESS) { *addr_ptr = result.virtual_protect.addr; *size_ptr = result.virtual_protect.size; if (old_prot) *old_prot = result.virtual_protect.prot; } return result.virtual_protect.status; } /* Fix the parameters */ size = ROUND_SIZE( addr, size ); base = ROUND_ADDR( addr, page_mask ); server_enter_uninterrupted_section( &csVirtual, &sigset ); if (!(view = VIRTUAL_FindView( base )) || (base + size > (char *)view->base + view->size)) { status = STATUS_INVALID_PARAMETER; } else { /* Make sure all the pages are committed */ p = view->prot + ((base - (char *)view->base) >> page_shift); prot = VIRTUAL_GetWin32Prot( *p ); for (i = size >> page_shift; i; i--, p++) { if (!(*p & VPROT_COMMITTED)) { status = STATUS_NOT_COMMITTED; break; } } if (!i) { if (old_prot) *old_prot = prot; vprot = VIRTUAL_GetProt( new_prot ) | VPROT_COMMITTED; if (!VIRTUAL_SetProt( view, base, size, vprot )) status = STATUS_ACCESS_DENIED; } } server_leave_uninterrupted_section( &csVirtual, &sigset ); if (status == STATUS_SUCCESS) { *addr_ptr = base; *size_ptr = size; } return status; } #define UNIMPLEMENTED_INFO_CLASS(c) \ case c: \ FIXME("(process=%p,addr=%p) Unimplemented information class: " #c "\n", process, addr); \ return STATUS_INVALID_INFO_CLASS /*********************************************************************** * NtQueryVirtualMemory (NTDLL.@) * ZwQueryVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtQueryVirtualMemory( HANDLE process, LPCVOID addr, MEMORY_INFORMATION_CLASS info_class, PVOID buffer, SIZE_T len, SIZE_T *res_len ) { FILE_VIEW *view; char *base, *alloc_base = 0; struct list *ptr; SIZE_T size = 0; MEMORY_BASIC_INFORMATION *info = buffer; sigset_t sigset; if (info_class != MemoryBasicInformation) { switch(info_class) { UNIMPLEMENTED_INFO_CLASS(MemoryWorkingSetList); UNIMPLEMENTED_INFO_CLASS(MemorySectionName); UNIMPLEMENTED_INFO_CLASS(MemoryBasicVlmInformation); default: FIXME("(%p,%p,info_class=%d,%p,%ld,%p) Unknown information class\n", process, addr, info_class, buffer, len, res_len); return STATUS_INVALID_INFO_CLASS; } } if (address_space_limit && addr >= address_space_limit) return STATUS_WORKING_SET_LIMIT_RANGE; if (process != NtCurrentProcess()) { NTSTATUS status; apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_query.type = APC_VIRTUAL_QUERY; call.virtual_query.addr = addr; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_query.status == STATUS_SUCCESS) { info->BaseAddress = result.virtual_query.base; info->AllocationBase = result.virtual_query.alloc_base; info->RegionSize = result.virtual_query.size; info->State = result.virtual_query.state; info->Protect = result.virtual_query.prot; info->AllocationProtect = result.virtual_query.alloc_prot; info->Type = result.virtual_query.alloc_type; if (res_len) *res_len = sizeof(*info); } return result.virtual_query.status; } base = ROUND_ADDR( addr, page_mask ); /* Find the view containing the address */ server_enter_uninterrupted_section( &csVirtual, &sigset ); ptr = list_head( &views_list ); for (;;) { if (!ptr) { /* make the address space end at the user limit, except if * the last view was mapped beyond that */ if (alloc_base <= (char *)user_space_limit) { if (user_space_limit && base >= (char *)user_space_limit) { server_leave_uninterrupted_section( &csVirtual, &sigset ); return STATUS_WORKING_SET_LIMIT_RANGE; } size = (char *)user_space_limit - alloc_base; } else size = (char *)address_space_limit - alloc_base; view = NULL; break; } view = LIST_ENTRY( ptr, struct file_view, entry ); if ((char *)view->base > base) { size = (char *)view->base - alloc_base; view = NULL; break; } if ((char *)view->base + view->size > base) { alloc_base = view->base; size = view->size; break; } alloc_base = (char *)view->base + view->size; ptr = list_next( &views_list, ptr ); } /* Fill the info structure */ if (!view) { info->State = MEM_FREE; info->Protect = PAGE_NOACCESS; info->AllocationBase = 0; info->AllocationProtect = 0; info->Type = 0; } else { BYTE vprot = view->prot[(base - alloc_base) >> page_shift]; info->State = (vprot & VPROT_COMMITTED) ? MEM_COMMIT : MEM_RESERVE; info->Protect = VIRTUAL_GetWin32Prot( vprot ); info->AllocationBase = alloc_base; info->AllocationProtect = VIRTUAL_GetWin32Prot( view->protect ); if (view->protect & VPROT_IMAGE) info->Type = MEM_IMAGE; else if (view->protect & VPROT_VALLOC) info->Type = MEM_PRIVATE; else info->Type = MEM_MAPPED; for (size = base - alloc_base; size < view->size; size += page_size) if (view->prot[size >> page_shift] != vprot) break; } server_leave_uninterrupted_section( &csVirtual, &sigset ); info->BaseAddress = base; info->RegionSize = size - (base - alloc_base); if (res_len) *res_len = sizeof(*info); return STATUS_SUCCESS; } /*********************************************************************** * NtLockVirtualMemory (NTDLL.@) * ZwLockVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtLockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown ) { NTSTATUS status = STATUS_SUCCESS; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_lock.type = APC_VIRTUAL_LOCK; call.virtual_lock.addr = *addr; call.virtual_lock.size = *size; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_lock.status == STATUS_SUCCESS) { *addr = result.virtual_lock.addr; *size = result.virtual_lock.size; } return result.virtual_lock.status; } *size = ROUND_SIZE( *addr, *size ); *addr = ROUND_ADDR( *addr, page_mask ); if (mlock( *addr, *size )) status = STATUS_ACCESS_DENIED; return status; } /*********************************************************************** * NtUnlockVirtualMemory (NTDLL.@) * ZwUnlockVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtUnlockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown ) { NTSTATUS status = STATUS_SUCCESS; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_unlock.type = APC_VIRTUAL_UNLOCK; call.virtual_unlock.addr = *addr; call.virtual_unlock.size = *size; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_unlock.status == STATUS_SUCCESS) { *addr = result.virtual_unlock.addr; *size = result.virtual_unlock.size; } return result.virtual_unlock.status; } *size = ROUND_SIZE( *addr, *size ); *addr = ROUND_ADDR( *addr, page_mask ); if (munlock( *addr, *size )) status = STATUS_ACCESS_DENIED; return status; } /*********************************************************************** * NtCreateSection (NTDLL.@) * ZwCreateSection (NTDLL.@) */ NTSTATUS WINAPI NtCreateSection( HANDLE *handle, ACCESS_MASK access, const OBJECT_ATTRIBUTES *attr, const LARGE_INTEGER *size, ULONG protect, ULONG sec_flags, HANDLE file ) { NTSTATUS ret; unsigned int vprot; DWORD len = (attr && attr->ObjectName) ? attr->ObjectName->Length : 0; struct security_descriptor *sd = NULL; struct object_attributes objattr; /* Check parameters */ if (len > MAX_PATH*sizeof(WCHAR)) return STATUS_NAME_TOO_LONG; objattr.rootdir = attr ? attr->RootDirectory : 0; objattr.sd_len = 0; objattr.name_len = len; if (attr) { ret = NTDLL_create_struct_sd( attr->SecurityDescriptor, &sd, &objattr.sd_len ); if (ret != STATUS_SUCCESS) return ret; } vprot = VIRTUAL_GetProt( protect ); if (sec_flags & SEC_RESERVE) { if (file) return STATUS_INVALID_PARAMETER; } else vprot |= VPROT_COMMITTED; if (sec_flags & SEC_NOCACHE) vprot |= VPROT_NOCACHE; if (sec_flags & SEC_IMAGE) vprot |= VPROT_IMAGE; /* Create the server object */ SERVER_START_REQ( create_mapping ) { req->access = access; req->attributes = (attr) ? attr->Attributes : 0; req->file_handle = file; req->size = size ? size->QuadPart : 0; req->protect = vprot; wine_server_add_data( req, &objattr, sizeof(objattr) ); if (objattr.sd_len) wine_server_add_data( req, sd, objattr.sd_len ); if (len) wine_server_add_data( req, attr->ObjectName->Buffer, len ); ret = wine_server_call( req ); *handle = reply->handle; } SERVER_END_REQ; NTDLL_free_struct_sd( sd ); return ret; } /*********************************************************************** * NtOpenSection (NTDLL.@) * ZwOpenSection (NTDLL.@) */ NTSTATUS WINAPI NtOpenSection( HANDLE *handle, ACCESS_MASK access, const OBJECT_ATTRIBUTES *attr ) { NTSTATUS ret; DWORD len = attr->ObjectName->Length; if (len > MAX_PATH*sizeof(WCHAR)) return STATUS_NAME_TOO_LONG; SERVER_START_REQ( open_mapping ) { req->access = access; req->attributes = attr->Attributes; req->rootdir = attr->RootDirectory; wine_server_add_data( req, attr->ObjectName->Buffer, len ); if (!(ret = wine_server_call( req ))) *handle = reply->handle; } SERVER_END_REQ; return ret; } /*********************************************************************** * NtMapViewOfSection (NTDLL.@) * ZwMapViewOfSection (NTDLL.@) */ NTSTATUS WINAPI NtMapViewOfSection( HANDLE handle, HANDLE process, PVOID *addr_ptr, ULONG zero_bits, SIZE_T commit_size, const LARGE_INTEGER *offset_ptr, SIZE_T *size_ptr, SECTION_INHERIT inherit, ULONG alloc_type, ULONG protect ) { NTSTATUS res; ULONGLONG full_size; SIZE_T size = 0; SIZE_T mask = get_mask( zero_bits ); int unix_handle = -1, needs_close; unsigned int prot; void *base; struct file_view *view; DWORD header_size; HANDLE dup_mapping, shared_file; LARGE_INTEGER offset; sigset_t sigset; offset.QuadPart = offset_ptr ? offset_ptr->QuadPart : 0; TRACE("handle=%p process=%p addr=%p off=%x%08x size=%lx access=%x\n", handle, process, *addr_ptr, offset.u.HighPart, offset.u.LowPart, size, protect ); /* Check parameters */ if ((offset.u.LowPart & mask) || (*addr_ptr && ((UINT_PTR)*addr_ptr & mask))) return STATUS_INVALID_PARAMETER; if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.map_view.type = APC_MAP_VIEW; call.map_view.handle = handle; call.map_view.addr = *addr_ptr; call.map_view.size = *size_ptr; call.map_view.offset = offset.QuadPart; call.map_view.zero_bits = zero_bits; call.map_view.alloc_type = alloc_type; call.map_view.prot = protect; res = NTDLL_queue_process_apc( process, &call, &result ); if (res != STATUS_SUCCESS) return res; if (result.map_view.status == STATUS_SUCCESS) { *addr_ptr = result.map_view.addr; *size_ptr = result.map_view.size; } return result.map_view.status; } SERVER_START_REQ( get_mapping_info ) { req->handle = handle; res = wine_server_call( req ); prot = reply->protect; base = reply->base; full_size = reply->size; header_size = reply->header_size; dup_mapping = reply->mapping; shared_file = reply->shared_file; } SERVER_END_REQ; if (res) return res; size = full_size; if (sizeof(size) < sizeof(full_size) && (size != full_size)) ERR( "Sizes larger than 4Gb (%x%08x) not supported on this platform\n", (DWORD)(full_size >> 32), (DWORD)full_size ); if ((res = server_get_unix_fd( handle, 0, &unix_handle, &needs_close, NULL, NULL ))) goto done; if (prot & VPROT_IMAGE) { if (shared_file) { int shared_fd, shared_needs_close; if ((res = server_get_unix_fd( shared_file, FILE_READ_DATA|FILE_WRITE_DATA, &shared_fd, &shared_needs_close, NULL, NULL ))) goto done; res = map_image( handle, unix_handle, base, size, mask, header_size, shared_fd, dup_mapping, addr_ptr ); if (shared_needs_close) close( shared_fd ); NtClose( shared_file ); } else { res = map_image( handle, unix_handle, base, size, mask, header_size, -1, dup_mapping, addr_ptr ); } if (needs_close) close( unix_handle ); if (!res) *size_ptr = size; return res; } if ((offset.QuadPart >= size) || (*size_ptr > size - offset.QuadPart)) { res = STATUS_INVALID_PARAMETER; goto done; } if (*size_ptr) size = ROUND_SIZE( offset.u.LowPart, *size_ptr ); else size = size - offset.QuadPart; switch(protect) { case PAGE_NOACCESS: break; case PAGE_READWRITE: case PAGE_EXECUTE_READWRITE: if (!(prot & VPROT_WRITE)) { res = STATUS_INVALID_PARAMETER; goto done; } /* fall through */ case PAGE_READONLY: case PAGE_WRITECOPY: case PAGE_EXECUTE: case PAGE_EXECUTE_READ: case PAGE_EXECUTE_WRITECOPY: if (prot & VPROT_READ) break; /* fall through */ default: res = STATUS_INVALID_PARAMETER; goto done; } /* FIXME: If a mapping is created with SEC_RESERVE and a process, * which has a view of this mapping commits some pages, they will * appear committed in all other processes, which have the same * view created. Since we don't support this yet, we create the * whole mapping committed. */ prot |= VPROT_COMMITTED; /* Reserve a properly aligned area */ server_enter_uninterrupted_section( &csVirtual, &sigset ); res = map_view( &view, *addr_ptr, size, mask, FALSE, prot ); if (res) { server_leave_uninterrupted_section( &csVirtual, &sigset ); goto done; } /* Map the file */ TRACE("handle=%p size=%lx offset=%x%08x\n", handle, size, offset.u.HighPart, offset.u.LowPart ); res = map_file_into_view( view, unix_handle, 0, size, offset.QuadPart, prot, !dup_mapping ); if (res == STATUS_SUCCESS) { *addr_ptr = view->base; *size_ptr = size; view->mapping = dup_mapping; dup_mapping = 0; /* don't close it */ } else { ERR( "map_file_into_view %p %lx %x%08x failed\n", view->base, size, offset.u.HighPart, offset.u.LowPart ); delete_view( view ); } server_leave_uninterrupted_section( &csVirtual, &sigset ); done: if (dup_mapping) NtClose( dup_mapping ); if (needs_close) close( unix_handle ); return res; } /*********************************************************************** * NtUnmapViewOfSection (NTDLL.@) * ZwUnmapViewOfSection (NTDLL.@) */ NTSTATUS WINAPI NtUnmapViewOfSection( HANDLE process, PVOID addr ) { FILE_VIEW *view; NTSTATUS status = STATUS_INVALID_PARAMETER; sigset_t sigset; void *base = ROUND_ADDR( addr, page_mask ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.unmap_view.type = APC_UNMAP_VIEW; call.unmap_view.addr = addr; status = NTDLL_queue_process_apc( process, &call, &result ); if (status == STATUS_SUCCESS) status = result.unmap_view.status; return status; } server_enter_uninterrupted_section( &csVirtual, &sigset ); if ((view = VIRTUAL_FindView( base )) && (base == view->base)) { delete_view( view ); status = STATUS_SUCCESS; } server_leave_uninterrupted_section( &csVirtual, &sigset ); return status; } /*********************************************************************** * NtFlushVirtualMemory (NTDLL.@) * ZwFlushVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtFlushVirtualMemory( HANDLE process, LPCVOID *addr_ptr, SIZE_T *size_ptr, ULONG unknown ) { FILE_VIEW *view; NTSTATUS status = STATUS_SUCCESS; sigset_t sigset; void *addr = ROUND_ADDR( *addr_ptr, page_mask ); if (process != NtCurrentProcess()) { apc_call_t call; apc_result_t result; memset( &call, 0, sizeof(call) ); call.virtual_flush.type = APC_VIRTUAL_FLUSH; call.virtual_flush.addr = addr; call.virtual_flush.size = *size_ptr; status = NTDLL_queue_process_apc( process, &call, &result ); if (status != STATUS_SUCCESS) return status; if (result.virtual_flush.status == STATUS_SUCCESS) { *addr_ptr = result.virtual_flush.addr; *size_ptr = result.virtual_flush.size; } return result.virtual_flush.status; } server_enter_uninterrupted_section( &csVirtual, &sigset ); if (!(view = VIRTUAL_FindView( addr ))) status = STATUS_INVALID_PARAMETER; else { if (!*size_ptr) *size_ptr = view->size; *addr_ptr = addr; if (msync( addr, *size_ptr, MS_SYNC )) status = STATUS_NOT_MAPPED_DATA; } server_leave_uninterrupted_section( &csVirtual, &sigset ); return status; } /*********************************************************************** * NtReadVirtualMemory (NTDLL.@) * ZwReadVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtReadVirtualMemory( HANDLE process, const void *addr, void *buffer, SIZE_T size, SIZE_T *bytes_read ) { NTSTATUS status; SERVER_START_REQ( read_process_memory ) { req->handle = process; req->addr = (void *)addr; wine_server_set_reply( req, buffer, size ); if ((status = wine_server_call( req ))) size = 0; } SERVER_END_REQ; if (bytes_read) *bytes_read = size; return status; } /*********************************************************************** * NtWriteVirtualMemory (NTDLL.@) * ZwWriteVirtualMemory (NTDLL.@) */ NTSTATUS WINAPI NtWriteVirtualMemory( HANDLE process, void *addr, const void *buffer, SIZE_T size, SIZE_T *bytes_written ) { NTSTATUS status; SERVER_START_REQ( write_process_memory ) { req->handle = process; req->addr = addr; wine_server_add_data( req, buffer, size ); if ((status = wine_server_call( req ))) size = 0; } SERVER_END_REQ; if (bytes_written) *bytes_written = size; return status; } /*********************************************************************** * NtAreMappedFilesTheSame (NTDLL.@) * ZwAreMappedFilesTheSame (NTDLL.@) */ NTSTATUS WINAPI NtAreMappedFilesTheSame(PVOID addr1, PVOID addr2) { TRACE("%p %p\n", addr1, addr2); return STATUS_NOT_SAME_DEVICE; }