2307 lines
76 KiB
C
2307 lines
76 KiB
C
/*
|
|
* Win32 virtual memory functions
|
|
*
|
|
* Copyright 1997, 2002 Alexandre Julliard
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#include "config.h"
|
|
#include "wine/port.h"
|
|
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#ifdef HAVE_SYS_ERRNO_H
|
|
#include <sys/errno.h>
|
|
#endif
|
|
#include <fcntl.h>
|
|
#ifdef HAVE_UNISTD_H
|
|
# include <unistd.h>
|
|
#endif
|
|
#include <stdarg.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
#include <string.h>
|
|
#include <sys/types.h>
|
|
#ifdef HAVE_SYS_STAT_H
|
|
# include <sys/stat.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_MMAN_H
|
|
# include <sys/mman.h>
|
|
#endif
|
|
#ifdef HAVE_VALGRIND_MEMCHECK_H
|
|
# include <valgrind/memcheck.h>
|
|
#endif
|
|
|
|
#define NONAMELESSUNION
|
|
#define NONAMELESSSTRUCT
|
|
#include "ntstatus.h"
|
|
#define WIN32_NO_STATUS
|
|
#include "windef.h"
|
|
#include "winternl.h"
|
|
#include "wine/library.h"
|
|
#include "wine/server.h"
|
|
#include "wine/list.h"
|
|
#include "wine/debug.h"
|
|
#include "ntdll_misc.h"
|
|
|
|
WINE_DEFAULT_DEBUG_CHANNEL(virtual);
|
|
WINE_DECLARE_DEBUG_CHANNEL(module);
|
|
|
|
#ifndef MS_SYNC
|
|
#define MS_SYNC 0
|
|
#endif
|
|
|
|
#ifndef MAP_NORESERVE
|
|
#define MAP_NORESERVE 0
|
|
#endif
|
|
|
|
/* File view */
|
|
typedef struct file_view
|
|
{
|
|
struct list entry; /* Entry in global view list */
|
|
void *base; /* Base address */
|
|
size_t size; /* Size in bytes */
|
|
HANDLE mapping; /* Handle to the file mapping */
|
|
BYTE flags; /* Allocation flags (VFLAG_*) */
|
|
BYTE protect; /* Protection for all pages at allocation time */
|
|
BYTE prot[1]; /* Protection byte for each page */
|
|
} FILE_VIEW;
|
|
|
|
/* Per-view flags */
|
|
#define VFLAG_SYSTEM 0x01 /* system view (underlying mmap not under our control) */
|
|
#define VFLAG_VALLOC 0x02 /* allocated by VirtualAlloc */
|
|
|
|
/* Conversion from VPROT_* to Win32 flags */
|
|
static const BYTE VIRTUAL_Win32Flags[16] =
|
|
{
|
|
PAGE_NOACCESS, /* 0 */
|
|
PAGE_READONLY, /* READ */
|
|
PAGE_READWRITE, /* WRITE */
|
|
PAGE_READWRITE, /* READ | WRITE */
|
|
PAGE_EXECUTE, /* EXEC */
|
|
PAGE_EXECUTE_READ, /* READ | EXEC */
|
|
PAGE_EXECUTE_READWRITE, /* WRITE | EXEC */
|
|
PAGE_EXECUTE_READWRITE, /* READ | WRITE | EXEC */
|
|
PAGE_WRITECOPY, /* WRITECOPY */
|
|
PAGE_WRITECOPY, /* READ | WRITECOPY */
|
|
PAGE_WRITECOPY, /* WRITE | WRITECOPY */
|
|
PAGE_WRITECOPY, /* READ | WRITE | WRITECOPY */
|
|
PAGE_EXECUTE_WRITECOPY, /* EXEC | WRITECOPY */
|
|
PAGE_EXECUTE_WRITECOPY, /* READ | EXEC | WRITECOPY */
|
|
PAGE_EXECUTE_WRITECOPY, /* WRITE | EXEC | WRITECOPY */
|
|
PAGE_EXECUTE_WRITECOPY /* READ | WRITE | EXEC | WRITECOPY */
|
|
};
|
|
|
|
static struct list views_list = LIST_INIT(views_list);
|
|
|
|
static RTL_CRITICAL_SECTION csVirtual;
|
|
static RTL_CRITICAL_SECTION_DEBUG critsect_debug =
|
|
{
|
|
0, 0, &csVirtual,
|
|
{ &critsect_debug.ProcessLocksList, &critsect_debug.ProcessLocksList },
|
|
0, 0, { (DWORD_PTR)(__FILE__ ": csVirtual") }
|
|
};
|
|
static RTL_CRITICAL_SECTION csVirtual = { &critsect_debug, -1, 0, 0, 0, 0 };
|
|
|
|
#ifdef __i386__
|
|
/* These are always the same on an i386, and it will be faster this way */
|
|
# define page_mask 0xfff
|
|
# define page_shift 12
|
|
# define page_size 0x1000
|
|
/* Note: these are Windows limits, you cannot change them. */
|
|
# define ADDRESS_SPACE_LIMIT ((void *)0xc0000000) /* top of the total available address space */
|
|
# define USER_SPACE_LIMIT ((void *)0x7fff0000) /* top of the user address space */
|
|
#else
|
|
static UINT page_shift;
|
|
static UINT page_size;
|
|
static UINT_PTR page_mask;
|
|
# define ADDRESS_SPACE_LIMIT 0 /* no limit needed on other platforms */
|
|
# define USER_SPACE_LIMIT 0 /* no limit needed on other platforms */
|
|
#endif /* __i386__ */
|
|
|
|
#define ROUND_ADDR(addr,mask) \
|
|
((void *)((UINT_PTR)(addr) & ~(UINT_PTR)(mask)))
|
|
|
|
#define ROUND_SIZE(addr,size) \
|
|
(((UINT)(size) + ((UINT_PTR)(addr) & page_mask) + page_mask) & ~page_mask)
|
|
|
|
#define VIRTUAL_DEBUG_DUMP_VIEW(view) \
|
|
do { if (TRACE_ON(virtual)) VIRTUAL_DumpView(view); } while (0)
|
|
|
|
static void *user_space_limit = USER_SPACE_LIMIT;
|
|
static void *preload_reserve_start;
|
|
static void *preload_reserve_end;
|
|
static int use_locks;
|
|
static int force_exec_prot; /* whether to force PROT_EXEC on all PROT_READ mmaps */
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_GetProtStr
|
|
*/
|
|
static const char *VIRTUAL_GetProtStr( BYTE prot )
|
|
{
|
|
static char buffer[6];
|
|
buffer[0] = (prot & VPROT_COMMITTED) ? 'c' : '-';
|
|
buffer[1] = (prot & VPROT_GUARD) ? 'g' : '-';
|
|
buffer[2] = (prot & VPROT_READ) ? 'r' : '-';
|
|
buffer[3] = (prot & VPROT_WRITECOPY) ? 'W' : ((prot & VPROT_WRITE) ? 'w' : '-');
|
|
buffer[4] = (prot & VPROT_EXEC) ? 'x' : '-';
|
|
buffer[5] = 0;
|
|
return buffer;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_GetUnixProt
|
|
*
|
|
* Convert page protections to protection for mmap/mprotect.
|
|
*/
|
|
static int VIRTUAL_GetUnixProt( BYTE vprot )
|
|
{
|
|
int prot = 0;
|
|
if ((vprot & VPROT_COMMITTED) && !(vprot & VPROT_GUARD))
|
|
{
|
|
if (vprot & VPROT_READ) prot |= PROT_READ;
|
|
if (vprot & VPROT_WRITE) prot |= PROT_WRITE;
|
|
if (vprot & VPROT_WRITECOPY) prot |= PROT_WRITE;
|
|
if (vprot & VPROT_EXEC) prot |= PROT_EXEC;
|
|
}
|
|
if (!prot) prot = PROT_NONE;
|
|
return prot;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_DumpView
|
|
*/
|
|
static void VIRTUAL_DumpView( FILE_VIEW *view )
|
|
{
|
|
UINT i, count;
|
|
char *addr = view->base;
|
|
BYTE prot = view->prot[0];
|
|
|
|
TRACE( "View: %p - %p", addr, addr + view->size - 1 );
|
|
if (view->flags & VFLAG_SYSTEM)
|
|
TRACE( " (system)\n" );
|
|
else if (view->flags & VFLAG_VALLOC)
|
|
TRACE( " (valloc)\n" );
|
|
else if (view->mapping)
|
|
TRACE( " %p\n", view->mapping );
|
|
else
|
|
TRACE( " (anonymous)\n");
|
|
|
|
for (count = i = 1; i < view->size >> page_shift; i++, count++)
|
|
{
|
|
if (view->prot[i] == prot) continue;
|
|
TRACE( " %p - %p %s\n",
|
|
addr, addr + (count << page_shift) - 1, VIRTUAL_GetProtStr(prot) );
|
|
addr += (count << page_shift);
|
|
prot = view->prot[i];
|
|
count = 0;
|
|
}
|
|
if (count)
|
|
TRACE( " %p - %p %s\n",
|
|
addr, addr + (count << page_shift) - 1, VIRTUAL_GetProtStr(prot) );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_Dump
|
|
*/
|
|
#if WINE_VM_DEBUG
|
|
static void VIRTUAL_Dump(void)
|
|
{
|
|
sigset_t sigset;
|
|
struct file_view *view;
|
|
|
|
TRACE( "Dump of all virtual memory views:\n" );
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
LIST_FOR_EACH_ENTRY( view, &views_list, FILE_VIEW, entry )
|
|
{
|
|
VIRTUAL_DumpView( view );
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
}
|
|
#endif
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_FindView
|
|
*
|
|
* Find the view containing a given address. The csVirtual section must be held by caller.
|
|
*
|
|
* PARAMS
|
|
* addr [I] Address
|
|
*
|
|
* RETURNS
|
|
* View: Success
|
|
* NULL: Failure
|
|
*/
|
|
static struct file_view *VIRTUAL_FindView( const void *addr )
|
|
{
|
|
struct file_view *view;
|
|
|
|
LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry )
|
|
{
|
|
if (view->base > addr) break;
|
|
if ((const char*)view->base + view->size > (const char*)addr) return view;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* get_mask
|
|
*/
|
|
static inline UINT_PTR get_mask( ULONG zero_bits )
|
|
{
|
|
if (!zero_bits) return 0xffff; /* allocations are aligned to 64K by default */
|
|
if (zero_bits < page_shift) zero_bits = page_shift;
|
|
return (1 << zero_bits) - 1;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* find_view_range
|
|
*
|
|
* Find the first view overlapping at least part of the specified range.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static struct file_view *find_view_range( const void *addr, size_t size )
|
|
{
|
|
struct file_view *view;
|
|
|
|
LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry )
|
|
{
|
|
if ((const char *)view->base >= (const char *)addr + size) break;
|
|
if ((const char *)view->base + view->size > (const char *)addr) return view;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* find_free_area
|
|
*
|
|
* Find a free area between views inside the specified range.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static void *find_free_area( void *base, void *end, size_t size, size_t mask, int top_down )
|
|
{
|
|
struct list *ptr;
|
|
void *start;
|
|
|
|
if (top_down)
|
|
{
|
|
start = ROUND_ADDR( (char *)end - size, mask );
|
|
if (start >= end || start < base) return NULL;
|
|
|
|
for (ptr = views_list.prev; ptr != &views_list; ptr = ptr->prev)
|
|
{
|
|
struct file_view *view = LIST_ENTRY( ptr, struct file_view, entry );
|
|
|
|
if ((char *)view->base + view->size <= (char *)start) break;
|
|
if ((char *)view->base >= (char *)start + size) continue;
|
|
start = ROUND_ADDR( (char *)view->base - size, mask );
|
|
/* stop if remaining space is not large enough */
|
|
if (!start || start >= end || start < base) return NULL;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
start = ROUND_ADDR( (char *)base + mask, mask );
|
|
if (start >= end || (char *)end - (char *)start < size) return NULL;
|
|
|
|
for (ptr = views_list.next; ptr != &views_list; ptr = ptr->next)
|
|
{
|
|
struct file_view *view = LIST_ENTRY( ptr, struct file_view, entry );
|
|
|
|
if ((char *)view->base >= (char *)start + size) break;
|
|
if ((char *)view->base + view->size <= (char *)start) continue;
|
|
start = ROUND_ADDR( (char *)view->base + view->size + mask, mask );
|
|
/* stop if remaining space is not large enough */
|
|
if (!start || start >= end || (char *)end - (char *)start < size) return NULL;
|
|
}
|
|
}
|
|
return start;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* add_reserved_area
|
|
*
|
|
* Add a reserved area to the list maintained by libwine.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static void add_reserved_area( void *addr, size_t size )
|
|
{
|
|
TRACE( "adding %p-%p\n", addr, (char *)addr + size );
|
|
|
|
if (addr < user_space_limit)
|
|
{
|
|
/* unmap the part of the area that is below the limit */
|
|
assert( (char *)addr + size > (char *)user_space_limit );
|
|
munmap( addr, (char *)user_space_limit - (char *)addr );
|
|
size -= (char *)user_space_limit - (char *)addr;
|
|
addr = user_space_limit;
|
|
}
|
|
/* blow away existing mappings */
|
|
wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED );
|
|
wine_mmap_add_reserved_area( addr, size );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* is_beyond_limit
|
|
*
|
|
* Check if an address range goes beyond a given limit.
|
|
*/
|
|
static inline int is_beyond_limit( const void *addr, size_t size, const void *limit )
|
|
{
|
|
return (limit && (addr >= limit || (const char *)addr + size > (const char *)limit));
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* unmap_area
|
|
*
|
|
* Unmap an area, or simply replace it by an empty mapping if it is
|
|
* in a reserved area. The csVirtual section must be held by caller.
|
|
*/
|
|
static inline void unmap_area( void *addr, size_t size )
|
|
{
|
|
if (wine_mmap_is_in_reserved_area( addr, size ))
|
|
wine_anon_mmap( addr, size, PROT_NONE, MAP_NORESERVE | MAP_FIXED );
|
|
else if (is_beyond_limit( addr, size, user_space_limit ))
|
|
add_reserved_area( addr, size );
|
|
else
|
|
munmap( addr, size );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* delete_view
|
|
*
|
|
* Deletes a view. The csVirtual section must be held by caller.
|
|
*/
|
|
static void delete_view( struct file_view *view ) /* [in] View */
|
|
{
|
|
if (!(view->flags & VFLAG_SYSTEM)) unmap_area( view->base, view->size );
|
|
list_remove( &view->entry );
|
|
if (view->mapping) NtClose( view->mapping );
|
|
free( view );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* create_view
|
|
*
|
|
* Create a view. The csVirtual section must be held by caller.
|
|
*/
|
|
static NTSTATUS create_view( struct file_view **view_ret, void *base, size_t size, BYTE vprot )
|
|
{
|
|
struct file_view *view;
|
|
struct list *ptr;
|
|
int unix_prot = VIRTUAL_GetUnixProt( vprot );
|
|
|
|
assert( !((UINT_PTR)base & page_mask) );
|
|
assert( !(size & page_mask) );
|
|
|
|
/* Create the view structure */
|
|
|
|
if (!(view = malloc( sizeof(*view) + (size >> page_shift) - 1 ))) return STATUS_NO_MEMORY;
|
|
|
|
view->base = base;
|
|
view->size = size;
|
|
view->flags = 0;
|
|
view->mapping = 0;
|
|
view->protect = vprot;
|
|
memset( view->prot, vprot & ~VPROT_IMAGE, size >> page_shift );
|
|
|
|
/* Insert it in the linked list */
|
|
|
|
LIST_FOR_EACH( ptr, &views_list )
|
|
{
|
|
struct file_view *next = LIST_ENTRY( ptr, struct file_view, entry );
|
|
if (next->base > base) break;
|
|
}
|
|
list_add_before( ptr, &view->entry );
|
|
|
|
/* Check for overlapping views. This can happen if the previous view
|
|
* was a system view that got unmapped behind our back. In that case
|
|
* we recover by simply deleting it. */
|
|
|
|
if ((ptr = list_prev( &views_list, &view->entry )) != NULL)
|
|
{
|
|
struct file_view *prev = LIST_ENTRY( ptr, struct file_view, entry );
|
|
if ((char *)prev->base + prev->size > (char *)base)
|
|
{
|
|
TRACE( "overlapping prev view %p-%p for %p-%p\n",
|
|
prev->base, (char *)prev->base + prev->size,
|
|
base, (char *)base + view->size );
|
|
assert( prev->flags & VFLAG_SYSTEM );
|
|
delete_view( prev );
|
|
}
|
|
}
|
|
if ((ptr = list_next( &views_list, &view->entry )) != NULL)
|
|
{
|
|
struct file_view *next = LIST_ENTRY( ptr, struct file_view, entry );
|
|
if ((char *)base + view->size > (char *)next->base)
|
|
{
|
|
TRACE( "overlapping next view %p-%p for %p-%p\n",
|
|
next->base, (char *)next->base + next->size,
|
|
base, (char *)base + view->size );
|
|
assert( next->flags & VFLAG_SYSTEM );
|
|
delete_view( next );
|
|
}
|
|
}
|
|
|
|
*view_ret = view;
|
|
VIRTUAL_DEBUG_DUMP_VIEW( view );
|
|
|
|
if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC))
|
|
{
|
|
TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 );
|
|
mprotect( base, size, unix_prot | PROT_EXEC );
|
|
}
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_GetWin32Prot
|
|
*
|
|
* Convert page protections to Win32 flags.
|
|
*/
|
|
static DWORD VIRTUAL_GetWin32Prot( BYTE vprot )
|
|
{
|
|
DWORD ret = VIRTUAL_Win32Flags[vprot & 0x0f];
|
|
if (vprot & VPROT_NOCACHE) ret |= PAGE_NOCACHE;
|
|
if (vprot & VPROT_GUARD) ret |= PAGE_GUARD;
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_GetProt
|
|
*
|
|
* Build page protections from Win32 flags.
|
|
*
|
|
* PARAMS
|
|
* protect [I] Win32 protection flags
|
|
*
|
|
* RETURNS
|
|
* Value of page protection flags
|
|
*/
|
|
static BYTE VIRTUAL_GetProt( DWORD protect )
|
|
{
|
|
BYTE vprot;
|
|
|
|
switch(protect & 0xff)
|
|
{
|
|
case PAGE_READONLY:
|
|
vprot = VPROT_READ;
|
|
break;
|
|
case PAGE_READWRITE:
|
|
vprot = VPROT_READ | VPROT_WRITE;
|
|
break;
|
|
case PAGE_WRITECOPY:
|
|
/* MSDN CreateFileMapping() states that if PAGE_WRITECOPY is given,
|
|
* that the hFile must have been opened with GENERIC_READ and
|
|
* GENERIC_WRITE access. This is WRONG as tests show that you
|
|
* only need GENERIC_READ access (at least for Win9x,
|
|
* FIXME: what about NT?). Thus, we don't put VPROT_WRITE in
|
|
* PAGE_WRITECOPY and PAGE_EXECUTE_WRITECOPY.
|
|
*/
|
|
vprot = VPROT_READ | VPROT_WRITECOPY;
|
|
break;
|
|
case PAGE_EXECUTE:
|
|
vprot = VPROT_EXEC;
|
|
break;
|
|
case PAGE_EXECUTE_READ:
|
|
vprot = VPROT_EXEC | VPROT_READ;
|
|
break;
|
|
case PAGE_EXECUTE_READWRITE:
|
|
vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITE;
|
|
break;
|
|
case PAGE_EXECUTE_WRITECOPY:
|
|
/* See comment for PAGE_WRITECOPY above */
|
|
vprot = VPROT_EXEC | VPROT_READ | VPROT_WRITECOPY;
|
|
break;
|
|
case PAGE_NOACCESS:
|
|
default:
|
|
vprot = 0;
|
|
break;
|
|
}
|
|
if (protect & PAGE_GUARD) vprot |= VPROT_GUARD;
|
|
if (protect & PAGE_NOCACHE) vprot |= VPROT_NOCACHE;
|
|
return vprot;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_SetProt
|
|
*
|
|
* Change the protection of a range of pages.
|
|
*
|
|
* RETURNS
|
|
* TRUE: Success
|
|
* FALSE: Failure
|
|
*/
|
|
static BOOL VIRTUAL_SetProt( FILE_VIEW *view, /* [in] Pointer to view */
|
|
void *base, /* [in] Starting address */
|
|
size_t size, /* [in] Size in bytes */
|
|
BYTE vprot ) /* [in] Protections to use */
|
|
{
|
|
int unix_prot = VIRTUAL_GetUnixProt(vprot);
|
|
|
|
TRACE("%p-%p %s\n",
|
|
base, (char *)base + size - 1, VIRTUAL_GetProtStr( vprot ) );
|
|
|
|
/* if setting stack guard pages, store the permissions first, as the guard may be
|
|
* triggered at any point after mprotect and change the permissions again */
|
|
if ((vprot & VPROT_GUARD) &&
|
|
((char *)base >= (char *)NtCurrentTeb()->DeallocationStack) &&
|
|
((char *)base < (char *)NtCurrentTeb()->Tib.StackBase))
|
|
{
|
|
memset( view->prot + (((char *)base - (char *)view->base) >> page_shift),
|
|
vprot, size >> page_shift );
|
|
mprotect( base, size, unix_prot );
|
|
VIRTUAL_DEBUG_DUMP_VIEW( view );
|
|
return TRUE;
|
|
}
|
|
|
|
if (force_exec_prot && (unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC))
|
|
{
|
|
TRACE( "forcing exec permission on %p-%p\n", base, (char *)base + size - 1 );
|
|
if (!mprotect( base, size, unix_prot | PROT_EXEC )) goto done;
|
|
/* exec + write may legitimately fail, in that case fall back to write only */
|
|
if (!(unix_prot & PROT_WRITE)) return FALSE;
|
|
}
|
|
|
|
if (mprotect( base, size, unix_prot )) return FALSE; /* FIXME: last error */
|
|
|
|
done:
|
|
memset( view->prot + (((char *)base - (char *)view->base) >> page_shift),
|
|
vprot, size >> page_shift );
|
|
VIRTUAL_DEBUG_DUMP_VIEW( view );
|
|
return TRUE;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* unmap_extra_space
|
|
*
|
|
* Release the extra memory while keeping the range starting on the granularity boundary.
|
|
*/
|
|
static inline void *unmap_extra_space( void *ptr, size_t total_size, size_t wanted_size, size_t mask )
|
|
{
|
|
if ((ULONG_PTR)ptr & mask)
|
|
{
|
|
size_t extra = mask + 1 - ((ULONG_PTR)ptr & mask);
|
|
munmap( ptr, extra );
|
|
ptr = (char *)ptr + extra;
|
|
total_size -= extra;
|
|
}
|
|
if (total_size > wanted_size)
|
|
munmap( (char *)ptr + wanted_size, total_size - wanted_size );
|
|
return ptr;
|
|
}
|
|
|
|
|
|
struct alloc_area
|
|
{
|
|
size_t size;
|
|
size_t mask;
|
|
int top_down;
|
|
void *result;
|
|
};
|
|
|
|
/***********************************************************************
|
|
* alloc_reserved_area_callback
|
|
*
|
|
* Try to map some space inside a reserved area. Callback for wine_mmap_enum_reserved_areas.
|
|
*/
|
|
static int alloc_reserved_area_callback( void *start, size_t size, void *arg )
|
|
{
|
|
static void * const address_space_start = (void *)0x110000;
|
|
struct alloc_area *alloc = arg;
|
|
void *end = (char *)start + size;
|
|
|
|
if (start < address_space_start) start = address_space_start;
|
|
if (user_space_limit && end > user_space_limit) end = user_space_limit;
|
|
if (start >= end) return 0;
|
|
|
|
/* make sure we don't touch the preloader reserved range */
|
|
if (preload_reserve_end >= start)
|
|
{
|
|
if (preload_reserve_end >= end)
|
|
{
|
|
if (preload_reserve_start <= start) return 0; /* no space in that area */
|
|
if (preload_reserve_start < end) end = preload_reserve_start;
|
|
}
|
|
else if (preload_reserve_start <= start) start = preload_reserve_end;
|
|
else
|
|
{
|
|
/* range is split in two by the preloader reservation, try first part */
|
|
if ((alloc->result = find_free_area( start, preload_reserve_start, alloc->size,
|
|
alloc->mask, alloc->top_down )))
|
|
return 1;
|
|
/* then fall through to try second part */
|
|
start = preload_reserve_end;
|
|
}
|
|
}
|
|
if ((alloc->result = find_free_area( start, end, alloc->size, alloc->mask, alloc->top_down )))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* map_view
|
|
*
|
|
* Create a view and mmap the corresponding memory area.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static NTSTATUS map_view( struct file_view **view_ret, void *base, size_t size, size_t mask,
|
|
int top_down, BYTE vprot )
|
|
{
|
|
void *ptr;
|
|
NTSTATUS status;
|
|
|
|
if (base)
|
|
{
|
|
if (is_beyond_limit( base, size, ADDRESS_SPACE_LIMIT ))
|
|
return STATUS_WORKING_SET_LIMIT_RANGE;
|
|
|
|
switch (wine_mmap_is_in_reserved_area( base, size ))
|
|
{
|
|
case -1: /* partially in a reserved area */
|
|
return STATUS_CONFLICTING_ADDRESSES;
|
|
|
|
case 0: /* not in a reserved area, do a normal allocation */
|
|
if ((ptr = wine_anon_mmap( base, size, VIRTUAL_GetUnixProt(vprot), 0 )) == (void *)-1)
|
|
{
|
|
if (errno == ENOMEM) return STATUS_NO_MEMORY;
|
|
return STATUS_INVALID_PARAMETER;
|
|
}
|
|
if (ptr != base)
|
|
{
|
|
/* We couldn't get the address we wanted */
|
|
if (is_beyond_limit( ptr, size, user_space_limit )) add_reserved_area( ptr, size );
|
|
else munmap( ptr, size );
|
|
return STATUS_CONFLICTING_ADDRESSES;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
case 1: /* in a reserved area, make sure the address is available */
|
|
if (find_view_range( base, size )) return STATUS_CONFLICTING_ADDRESSES;
|
|
/* replace the reserved area by our mapping */
|
|
if ((ptr = wine_anon_mmap( base, size, VIRTUAL_GetUnixProt(vprot), MAP_FIXED )) != base)
|
|
return STATUS_INVALID_PARAMETER;
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
size_t view_size = size + mask + 1;
|
|
struct alloc_area alloc;
|
|
|
|
alloc.size = size;
|
|
alloc.mask = mask;
|
|
alloc.top_down = top_down;
|
|
if (wine_mmap_enum_reserved_areas( alloc_reserved_area_callback, &alloc, top_down ))
|
|
{
|
|
ptr = alloc.result;
|
|
TRACE( "got mem in reserved area %p-%p\n", ptr, (char *)ptr + size );
|
|
if (wine_anon_mmap( ptr, size, VIRTUAL_GetUnixProt(vprot), MAP_FIXED ) != ptr)
|
|
return STATUS_INVALID_PARAMETER;
|
|
goto done;
|
|
}
|
|
|
|
for (;;)
|
|
{
|
|
if ((ptr = wine_anon_mmap( NULL, view_size, VIRTUAL_GetUnixProt(vprot), 0 )) == (void *)-1)
|
|
{
|
|
if (errno == ENOMEM) return STATUS_NO_MEMORY;
|
|
return STATUS_INVALID_PARAMETER;
|
|
}
|
|
TRACE( "got mem with anon mmap %p-%p\n", ptr, (char *)ptr + size );
|
|
/* if we got something beyond the user limit, unmap it and retry */
|
|
if (is_beyond_limit( ptr, view_size, user_space_limit )) add_reserved_area( ptr, view_size );
|
|
else break;
|
|
}
|
|
ptr = unmap_extra_space( ptr, view_size, size, mask );
|
|
}
|
|
done:
|
|
status = create_view( view_ret, ptr, size, vprot );
|
|
if (status != STATUS_SUCCESS) unmap_area( ptr, size );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* unaligned_mmap
|
|
*
|
|
* Linux kernels before 2.4.x can support non page-aligned offsets, as
|
|
* long as the offset is aligned to the filesystem block size. This is
|
|
* a big performance gain so we want to take advantage of it.
|
|
*
|
|
* However, when we use 64-bit file support this doesn't work because
|
|
* glibc rejects unaligned offsets. Also glibc 2.1.3 mmap64 is broken
|
|
* in that it rounds unaligned offsets down to a page boundary. For
|
|
* these reasons we do a direct system call here.
|
|
*/
|
|
static void *unaligned_mmap( void *addr, size_t length, unsigned int prot,
|
|
unsigned int flags, int fd, off_t offset )
|
|
{
|
|
#if defined(linux) && defined(__i386__) && defined(__GNUC__)
|
|
if (!(offset >> 32) && (offset & page_mask))
|
|
{
|
|
int ret;
|
|
|
|
struct
|
|
{
|
|
void *addr;
|
|
unsigned int length;
|
|
unsigned int prot;
|
|
unsigned int flags;
|
|
unsigned int fd;
|
|
unsigned int offset;
|
|
} args;
|
|
|
|
args.addr = addr;
|
|
args.length = length;
|
|
args.prot = prot;
|
|
args.flags = flags;
|
|
args.fd = fd;
|
|
args.offset = offset;
|
|
|
|
__asm__ __volatile__("push %%ebx\n\t"
|
|
"movl %2,%%ebx\n\t"
|
|
"int $0x80\n\t"
|
|
"popl %%ebx"
|
|
: "=a" (ret)
|
|
: "0" (90), /* SYS_mmap */
|
|
"q" (&args)
|
|
: "memory" );
|
|
if (ret < 0 && ret > -4096)
|
|
{
|
|
errno = -ret;
|
|
ret = -1;
|
|
}
|
|
return (void *)ret;
|
|
}
|
|
#endif
|
|
return mmap( addr, length, prot, flags, fd, offset );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* map_file_into_view
|
|
*
|
|
* Wrapper for mmap() to map a file into a view, falling back to read if mmap fails.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static NTSTATUS map_file_into_view( struct file_view *view, int fd, size_t start, size_t size,
|
|
off_t offset, BYTE vprot, BOOL removable )
|
|
{
|
|
void *ptr;
|
|
int prot = VIRTUAL_GetUnixProt( vprot );
|
|
BOOL shared_write = (vprot & VPROT_WRITE) != 0;
|
|
|
|
assert( start < view->size );
|
|
assert( start + size <= view->size );
|
|
|
|
/* only try mmap if media is not removable (or if we require write access) */
|
|
if (!removable || shared_write)
|
|
{
|
|
int flags = MAP_FIXED | (shared_write ? MAP_SHARED : MAP_PRIVATE);
|
|
|
|
if (unaligned_mmap( (char *)view->base + start, size, prot, flags, fd, offset ) != (void *)-1)
|
|
goto done;
|
|
|
|
/* mmap() failed; if this is because the file offset is not */
|
|
/* page-aligned (EINVAL), or because the underlying filesystem */
|
|
/* does not support mmap() (ENOEXEC,ENODEV), we do it by hand. */
|
|
if ((errno != ENOEXEC) && (errno != EINVAL) && (errno != ENODEV)) return FILE_GetNtStatus();
|
|
if (shared_write) /* we cannot fake shared write mappings */
|
|
{
|
|
if (errno == EINVAL) return STATUS_INVALID_PARAMETER;
|
|
ERR( "shared writable mmap not supported, broken filesystem?\n" );
|
|
return STATUS_NOT_SUPPORTED;
|
|
}
|
|
}
|
|
|
|
/* Reserve the memory with an anonymous mmap */
|
|
ptr = wine_anon_mmap( (char *)view->base + start, size, PROT_READ | PROT_WRITE, MAP_FIXED );
|
|
if (ptr == (void *)-1) return FILE_GetNtStatus();
|
|
/* Now read in the file */
|
|
pread( fd, ptr, size, offset );
|
|
if (prot != (PROT_READ|PROT_WRITE)) mprotect( ptr, size, prot ); /* Set the right protection */
|
|
done:
|
|
memset( view->prot + (start >> page_shift), vprot, ROUND_SIZE(start,size) >> page_shift );
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* decommit_view
|
|
*
|
|
* Decommit some pages of a given view.
|
|
* The csVirtual section must be held by caller.
|
|
*/
|
|
static NTSTATUS decommit_pages( struct file_view *view, size_t start, size_t size )
|
|
{
|
|
if (wine_anon_mmap( (char *)view->base + start, size, PROT_NONE, MAP_FIXED ) != (void *)-1)
|
|
{
|
|
BYTE *p = view->prot + (start >> page_shift);
|
|
size >>= page_shift;
|
|
while (size--) *p++ &= ~VPROT_COMMITTED;
|
|
return STATUS_SUCCESS;
|
|
}
|
|
return FILE_GetNtStatus();
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* map_image
|
|
*
|
|
* Map an executable (PE format) image into memory.
|
|
*/
|
|
static NTSTATUS map_image( HANDLE hmapping, int fd, char *base, SIZE_T total_size, SIZE_T mask,
|
|
SIZE_T header_size, int shared_fd, HANDLE dup_mapping, PVOID *addr_ptr )
|
|
{
|
|
IMAGE_DOS_HEADER *dos;
|
|
IMAGE_NT_HEADERS *nt;
|
|
IMAGE_SECTION_HEADER *sec;
|
|
IMAGE_DATA_DIRECTORY *imports;
|
|
NTSTATUS status = STATUS_CONFLICTING_ADDRESSES;
|
|
int i;
|
|
off_t pos;
|
|
sigset_t sigset;
|
|
struct stat st;
|
|
struct file_view *view = NULL;
|
|
char *ptr, *header_end;
|
|
|
|
/* zero-map the whole range */
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (base >= (char *)0x110000) /* make sure the DOS area remains free */
|
|
status = map_view( &view, base, total_size, mask, FALSE,
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY | VPROT_IMAGE );
|
|
|
|
if (status == STATUS_CONFLICTING_ADDRESSES)
|
|
status = map_view( &view, NULL, total_size, mask, FALSE,
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_EXEC | VPROT_WRITECOPY | VPROT_IMAGE );
|
|
|
|
if (status != STATUS_SUCCESS) goto error;
|
|
|
|
ptr = view->base;
|
|
TRACE_(module)( "mapped PE file at %p-%p\n", ptr, ptr + total_size );
|
|
|
|
/* map the header */
|
|
|
|
if (fstat( fd, &st ) == -1)
|
|
{
|
|
status = FILE_GetNtStatus();
|
|
goto error;
|
|
}
|
|
status = STATUS_INVALID_IMAGE_FORMAT; /* generic error */
|
|
if (!st.st_size) goto error;
|
|
header_size = min( header_size, st.st_size );
|
|
if (map_file_into_view( view, fd, 0, header_size, 0, VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY,
|
|
!dup_mapping ) != STATUS_SUCCESS) goto error;
|
|
dos = (IMAGE_DOS_HEADER *)ptr;
|
|
nt = (IMAGE_NT_HEADERS *)(ptr + dos->e_lfanew);
|
|
header_end = ptr + ROUND_SIZE( 0, header_size );
|
|
memset( ptr + header_size, 0, header_end - (ptr + header_size) );
|
|
if ((char *)(nt + 1) > header_end) goto error;
|
|
sec = (IMAGE_SECTION_HEADER*)((char*)&nt->OptionalHeader+nt->FileHeader.SizeOfOptionalHeader);
|
|
if ((char *)(sec + nt->FileHeader.NumberOfSections) > header_end) goto error;
|
|
|
|
imports = nt->OptionalHeader.DataDirectory + IMAGE_DIRECTORY_ENTRY_IMPORT;
|
|
if (!imports->Size || !imports->VirtualAddress) imports = NULL;
|
|
|
|
/* check the architecture */
|
|
|
|
if (nt->FileHeader.Machine != IMAGE_FILE_MACHINE_I386)
|
|
{
|
|
MESSAGE("Trying to load PE image for unsupported architecture (");
|
|
switch (nt->FileHeader.Machine)
|
|
{
|
|
case IMAGE_FILE_MACHINE_UNKNOWN: MESSAGE("Unknown"); break;
|
|
case IMAGE_FILE_MACHINE_I860: MESSAGE("I860"); break;
|
|
case IMAGE_FILE_MACHINE_R3000: MESSAGE("R3000"); break;
|
|
case IMAGE_FILE_MACHINE_R4000: MESSAGE("R4000"); break;
|
|
case IMAGE_FILE_MACHINE_R10000: MESSAGE("R10000"); break;
|
|
case IMAGE_FILE_MACHINE_ALPHA: MESSAGE("Alpha"); break;
|
|
case IMAGE_FILE_MACHINE_POWERPC: MESSAGE("PowerPC"); break;
|
|
case IMAGE_FILE_MACHINE_IA64: MESSAGE("IA-64"); break;
|
|
case IMAGE_FILE_MACHINE_ALPHA64: MESSAGE("Alpha-64"); break;
|
|
case IMAGE_FILE_MACHINE_AMD64: MESSAGE("AMD-64"); break;
|
|
case IMAGE_FILE_MACHINE_ARM: MESSAGE("ARM"); break;
|
|
default: MESSAGE("Unknown-%04x", nt->FileHeader.Machine); break;
|
|
}
|
|
MESSAGE(")\n");
|
|
goto error;
|
|
}
|
|
|
|
/* check for non page-aligned binary */
|
|
|
|
if (nt->OptionalHeader.SectionAlignment <= page_mask)
|
|
{
|
|
/* unaligned sections, this happens for native subsystem binaries */
|
|
/* in that case Windows simply maps in the whole file */
|
|
|
|
if (map_file_into_view( view, fd, 0, total_size, 0, VPROT_COMMITTED | VPROT_READ,
|
|
!dup_mapping ) != STATUS_SUCCESS) goto error;
|
|
|
|
/* check that all sections are loaded at the right offset */
|
|
if (nt->OptionalHeader.FileAlignment != nt->OptionalHeader.SectionAlignment) goto error;
|
|
for (i = 0; i < nt->FileHeader.NumberOfSections; i++)
|
|
{
|
|
if (sec[i].VirtualAddress != sec[i].PointerToRawData)
|
|
goto error; /* Windows refuses to load in that case too */
|
|
}
|
|
|
|
/* set the image protections */
|
|
VIRTUAL_SetProt( view, ptr, total_size,
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY | VPROT_EXEC );
|
|
|
|
/* no relocations are performed on non page-aligned binaries */
|
|
goto done;
|
|
}
|
|
|
|
|
|
/* map all the sections */
|
|
|
|
for (i = pos = 0; i < nt->FileHeader.NumberOfSections; i++, sec++)
|
|
{
|
|
static const SIZE_T sector_align = 0x1ff;
|
|
SIZE_T map_size, file_start, file_size, end;
|
|
|
|
if (!sec->Misc.VirtualSize)
|
|
map_size = ROUND_SIZE( 0, sec->SizeOfRawData );
|
|
else
|
|
map_size = ROUND_SIZE( 0, sec->Misc.VirtualSize );
|
|
|
|
/* file positions are rounded to sector boundaries regardless of OptionalHeader.FileAlignment */
|
|
file_start = sec->PointerToRawData & ~sector_align;
|
|
file_size = (sec->SizeOfRawData + (sec->PointerToRawData & sector_align) + sector_align) & ~sector_align;
|
|
if (file_size > map_size) file_size = map_size;
|
|
|
|
/* a few sanity checks */
|
|
end = sec->VirtualAddress + ROUND_SIZE( sec->VirtualAddress, map_size );
|
|
if (sec->VirtualAddress > total_size || end > total_size || end < sec->VirtualAddress)
|
|
{
|
|
WARN_(module)( "Section %.8s too large (%x+%lx/%lx)\n",
|
|
sec->Name, sec->VirtualAddress, map_size, total_size );
|
|
goto error;
|
|
}
|
|
|
|
if ((sec->Characteristics & IMAGE_SCN_MEM_SHARED) &&
|
|
(sec->Characteristics & IMAGE_SCN_MEM_WRITE))
|
|
{
|
|
TRACE_(module)( "mapping shared section %.8s at %p off %x (%x) size %lx (%lx) flags %x\n",
|
|
sec->Name, ptr + sec->VirtualAddress,
|
|
sec->PointerToRawData, (int)pos, file_size, map_size,
|
|
sec->Characteristics );
|
|
if (map_file_into_view( view, shared_fd, sec->VirtualAddress, map_size, pos,
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_WRITE,
|
|
FALSE ) != STATUS_SUCCESS)
|
|
{
|
|
ERR_(module)( "Could not map shared section %.8s\n", sec->Name );
|
|
goto error;
|
|
}
|
|
|
|
/* check if the import directory falls inside this section */
|
|
if (imports && imports->VirtualAddress >= sec->VirtualAddress &&
|
|
imports->VirtualAddress < sec->VirtualAddress + map_size)
|
|
{
|
|
UINT_PTR base = imports->VirtualAddress & ~page_mask;
|
|
UINT_PTR end = base + ROUND_SIZE( imports->VirtualAddress, imports->Size );
|
|
if (end > sec->VirtualAddress + map_size) end = sec->VirtualAddress + map_size;
|
|
if (end > base)
|
|
map_file_into_view( view, shared_fd, base, end - base,
|
|
pos + (base - sec->VirtualAddress),
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY,
|
|
FALSE );
|
|
}
|
|
pos += map_size;
|
|
continue;
|
|
}
|
|
|
|
TRACE_(module)( "mapping section %.8s at %p off %x size %x virt %x flags %x\n",
|
|
sec->Name, ptr + sec->VirtualAddress,
|
|
sec->PointerToRawData, sec->SizeOfRawData,
|
|
sec->Misc.VirtualSize, sec->Characteristics );
|
|
|
|
if (!sec->PointerToRawData || !file_size) continue;
|
|
|
|
/* Note: if the section is not aligned properly map_file_into_view will magically
|
|
* fall back to read(), so we don't need to check anything here.
|
|
*/
|
|
end = file_start + file_size;
|
|
if (sec->PointerToRawData >= st.st_size ||
|
|
end > ((st.st_size + sector_align) & ~sector_align) ||
|
|
end < file_start ||
|
|
map_file_into_view( view, fd, sec->VirtualAddress, file_size, file_start,
|
|
VPROT_COMMITTED | VPROT_READ | VPROT_WRITECOPY,
|
|
!dup_mapping ) != STATUS_SUCCESS)
|
|
{
|
|
ERR_(module)( "Could not map section %.8s, file probably truncated\n", sec->Name );
|
|
goto error;
|
|
}
|
|
|
|
if (file_size & page_mask)
|
|
{
|
|
end = ROUND_SIZE( 0, file_size );
|
|
if (end > map_size) end = map_size;
|
|
TRACE_(module)("clearing %p - %p\n",
|
|
ptr + sec->VirtualAddress + file_size,
|
|
ptr + sec->VirtualAddress + end );
|
|
memset( ptr + sec->VirtualAddress + file_size, 0, end - file_size );
|
|
}
|
|
}
|
|
|
|
|
|
/* perform base relocation, if necessary */
|
|
|
|
if (ptr != base &&
|
|
((nt->FileHeader.Characteristics & IMAGE_FILE_DLL) ||
|
|
!NtCurrentTeb()->Peb->ImageBaseAddress) )
|
|
{
|
|
int delta = ptr - base;
|
|
IMAGE_BASE_RELOCATION *rel, *end;
|
|
const IMAGE_DATA_DIRECTORY *relocs;
|
|
|
|
if (nt->FileHeader.Characteristics & IMAGE_FILE_RELOCS_STRIPPED)
|
|
{
|
|
WARN_(module)( "Need to relocate module from %p to %p, but there are no relocation records\n",
|
|
base, ptr );
|
|
status = STATUS_CONFLICTING_ADDRESSES;
|
|
goto error;
|
|
}
|
|
|
|
TRACE_(module)( "relocating from %p-%p to %p-%p\n",
|
|
base, base + total_size, ptr, ptr + total_size );
|
|
|
|
relocs = &nt->OptionalHeader.DataDirectory[IMAGE_DIRECTORY_ENTRY_BASERELOC];
|
|
rel = (IMAGE_BASE_RELOCATION *)(ptr + relocs->VirtualAddress);
|
|
end = (IMAGE_BASE_RELOCATION *)(ptr + relocs->VirtualAddress + relocs->Size);
|
|
|
|
while (rel <= end - 1 && rel->SizeOfBlock)
|
|
{
|
|
if (rel->VirtualAddress >= total_size)
|
|
{
|
|
WARN_(module)( "invalid address %p in relocation %p\n", ptr + rel->VirtualAddress, rel );
|
|
status = STATUS_ACCESS_VIOLATION;
|
|
goto error;
|
|
}
|
|
rel = LdrProcessRelocationBlock( ptr + rel->VirtualAddress,
|
|
(rel->SizeOfBlock - sizeof(*rel)) / sizeof(USHORT),
|
|
(USHORT *)(rel + 1), delta );
|
|
if (!rel) goto error;
|
|
}
|
|
}
|
|
|
|
/* set the image protections */
|
|
|
|
VIRTUAL_SetProt( view, ptr, ROUND_SIZE( 0, header_size ), VPROT_COMMITTED | VPROT_READ );
|
|
|
|
sec = (IMAGE_SECTION_HEADER*)((char *)&nt->OptionalHeader+nt->FileHeader.SizeOfOptionalHeader);
|
|
for (i = 0; i < nt->FileHeader.NumberOfSections; i++, sec++)
|
|
{
|
|
SIZE_T size;
|
|
BYTE vprot = VPROT_COMMITTED;
|
|
|
|
if (sec->Misc.VirtualSize)
|
|
size = ROUND_SIZE( sec->VirtualAddress, sec->Misc.VirtualSize );
|
|
else
|
|
size = ROUND_SIZE( sec->VirtualAddress, sec->SizeOfRawData );
|
|
|
|
if (sec->Characteristics & IMAGE_SCN_MEM_READ) vprot |= VPROT_READ;
|
|
if (sec->Characteristics & IMAGE_SCN_MEM_WRITE) vprot |= VPROT_READ|VPROT_WRITECOPY;
|
|
if (sec->Characteristics & IMAGE_SCN_MEM_EXECUTE) vprot |= VPROT_EXEC;
|
|
|
|
/* Dumb game crack lets the AOEP point into a data section. Adjust. */
|
|
if ((nt->OptionalHeader.AddressOfEntryPoint >= sec->VirtualAddress) &&
|
|
(nt->OptionalHeader.AddressOfEntryPoint < sec->VirtualAddress + size))
|
|
vprot |= VPROT_EXEC;
|
|
|
|
VIRTUAL_SetProt( view, ptr + sec->VirtualAddress, size, vprot );
|
|
}
|
|
|
|
done:
|
|
view->mapping = dup_mapping;
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
*addr_ptr = ptr;
|
|
return STATUS_SUCCESS;
|
|
|
|
error:
|
|
if (view) delete_view( view );
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
if (dup_mapping) NtClose( dup_mapping );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* virtual_init
|
|
*/
|
|
void virtual_init(void)
|
|
{
|
|
const char *preload;
|
|
#ifndef page_mask
|
|
page_size = getpagesize();
|
|
page_mask = page_size - 1;
|
|
/* Make sure we have a power of 2 */
|
|
assert( !(page_size & page_mask) );
|
|
page_shift = 0;
|
|
while ((1 << page_shift) != page_size) page_shift++;
|
|
#endif /* page_mask */
|
|
if ((preload = getenv("WINEPRELOADRESERVE")))
|
|
{
|
|
unsigned long start, end;
|
|
if (sscanf( preload, "%lx-%lx", &start, &end ) == 2)
|
|
{
|
|
preload_reserve_start = (void *)start;
|
|
preload_reserve_end = (void *)end;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* virtual_init_threading
|
|
*/
|
|
void virtual_init_threading(void)
|
|
{
|
|
use_locks = 1;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* virtual_alloc_thread_stack
|
|
*/
|
|
NTSTATUS virtual_alloc_thread_stack( void *base, SIZE_T size )
|
|
{
|
|
FILE_VIEW *view;
|
|
NTSTATUS status;
|
|
sigset_t sigset;
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (base) /* already allocated, create a system view */
|
|
{
|
|
size = ROUND_SIZE( base, size );
|
|
base = ROUND_ADDR( base, page_mask );
|
|
if ((status = create_view( &view, base, size,
|
|
VPROT_READ | VPROT_WRITE | VPROT_COMMITTED )) != STATUS_SUCCESS)
|
|
goto done;
|
|
view->flags |= VFLAG_VALLOC | VFLAG_SYSTEM;
|
|
}
|
|
else
|
|
{
|
|
size = (size + 0xffff) & ~0xffff; /* round to 64K boundary */
|
|
if ((status = map_view( &view, NULL, size, 0xffff, 0,
|
|
VPROT_READ | VPROT_WRITE | VPROT_COMMITTED )) != STATUS_SUCCESS)
|
|
goto done;
|
|
view->flags |= VFLAG_VALLOC;
|
|
#ifdef VALGRIND_STACK_REGISTER
|
|
/* no need to de-register the stack as it's the one of the main thread */
|
|
VALGRIND_STACK_REGISTER( view->base, (char *)view->base + view->size );
|
|
#endif
|
|
}
|
|
|
|
/* setup no access guard page */
|
|
VIRTUAL_SetProt( view, view->base, page_size, VPROT_COMMITTED );
|
|
VIRTUAL_SetProt( view, (char *)view->base + page_size, page_size,
|
|
VPROT_READ | VPROT_WRITE | VPROT_COMMITTED | VPROT_GUARD );
|
|
|
|
/* note: limit is lower than base since the stack grows down */
|
|
NtCurrentTeb()->DeallocationStack = view->base;
|
|
NtCurrentTeb()->Tib.StackBase = (char *)view->base + view->size;
|
|
NtCurrentTeb()->Tib.StackLimit = (char *)view->base + 2 * page_size;
|
|
|
|
done:
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* virtual_clear_thread_stack
|
|
*
|
|
* Clear the stack contents before calling the main entry point, some broken apps need that.
|
|
*/
|
|
void virtual_clear_thread_stack(void)
|
|
{
|
|
void *stack = NtCurrentTeb()->Tib.StackLimit;
|
|
size_t size = (char *)NtCurrentTeb()->Tib.StackBase - (char *)NtCurrentTeb()->Tib.StackLimit;
|
|
|
|
wine_anon_mmap( stack, size, PROT_READ | PROT_WRITE, MAP_FIXED );
|
|
if (force_exec_prot) mprotect( stack, size, PROT_READ | PROT_WRITE | PROT_EXEC );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_HandleFault
|
|
*/
|
|
NTSTATUS VIRTUAL_HandleFault( LPCVOID addr )
|
|
{
|
|
FILE_VIEW *view;
|
|
NTSTATUS ret = STATUS_ACCESS_VIOLATION;
|
|
sigset_t sigset;
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
if ((view = VIRTUAL_FindView( addr )))
|
|
{
|
|
void *page = ROUND_ADDR( addr, page_mask );
|
|
BYTE vprot = view->prot[((const char *)page - (const char *)view->base) >> page_shift];
|
|
if (vprot & VPROT_GUARD)
|
|
{
|
|
VIRTUAL_SetProt( view, page, page_size, vprot & ~VPROT_GUARD );
|
|
ret = STATUS_GUARD_PAGE_VIOLATION;
|
|
}
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return ret;
|
|
}
|
|
|
|
|
|
|
|
/***********************************************************************
|
|
* virtual_handle_stack_fault
|
|
*
|
|
* Handle an access fault inside the current thread stack.
|
|
* Called from inside a signal handler.
|
|
*/
|
|
BOOL virtual_handle_stack_fault( void *addr )
|
|
{
|
|
FILE_VIEW *view;
|
|
BOOL ret = FALSE;
|
|
|
|
RtlEnterCriticalSection( &csVirtual ); /* no need for signal masking inside signal handler */
|
|
if ((view = VIRTUAL_FindView( addr )))
|
|
{
|
|
void *page = ROUND_ADDR( addr, page_mask );
|
|
BYTE vprot = view->prot[((const char *)page - (const char *)view->base) >> page_shift];
|
|
if (vprot & VPROT_GUARD)
|
|
{
|
|
VIRTUAL_SetProt( view, page, page_size, vprot & ~VPROT_GUARD );
|
|
if ((char *)page + page_size == NtCurrentTeb()->Tib.StackLimit)
|
|
NtCurrentTeb()->Tib.StackLimit = page;
|
|
ret = TRUE;
|
|
}
|
|
}
|
|
RtlLeaveCriticalSection( &csVirtual );
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_SetForceExec
|
|
*
|
|
* Whether to force exec prot on all views.
|
|
*/
|
|
void VIRTUAL_SetForceExec( BOOL enable )
|
|
{
|
|
struct file_view *view;
|
|
sigset_t sigset;
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
if (!force_exec_prot != !enable) /* change all existing views */
|
|
{
|
|
force_exec_prot = enable;
|
|
|
|
LIST_FOR_EACH_ENTRY( view, &views_list, struct file_view, entry )
|
|
{
|
|
UINT i, count;
|
|
int unix_prot;
|
|
char *addr = view->base;
|
|
BYTE prot = view->prot[0];
|
|
|
|
for (count = i = 1; i < view->size >> page_shift; i++, count++)
|
|
{
|
|
if (view->prot[i] == prot) continue;
|
|
unix_prot = VIRTUAL_GetUnixProt( prot );
|
|
if ((unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC))
|
|
{
|
|
TRACE( "%s exec prot for %p-%p\n",
|
|
force_exec_prot ? "enabling" : "disabling",
|
|
addr, addr + (count << page_shift) - 1 );
|
|
mprotect( addr, count << page_shift,
|
|
unix_prot | (force_exec_prot ? PROT_EXEC : 0) );
|
|
}
|
|
addr += (count << page_shift);
|
|
prot = view->prot[i];
|
|
count = 0;
|
|
}
|
|
if (count)
|
|
{
|
|
unix_prot = VIRTUAL_GetUnixProt( prot );
|
|
if ((unix_prot & PROT_READ) && !(unix_prot & PROT_EXEC))
|
|
{
|
|
TRACE( "%s exec prot for %p-%p\n",
|
|
force_exec_prot ? "enabling" : "disabling",
|
|
addr, addr + (count << page_shift) - 1 );
|
|
mprotect( addr, count << page_shift,
|
|
unix_prot | (force_exec_prot ? PROT_EXEC : 0) );
|
|
}
|
|
}
|
|
}
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* VIRTUAL_UseLargeAddressSpace
|
|
*
|
|
* Increase the address space size for apps that support it.
|
|
*/
|
|
void VIRTUAL_UseLargeAddressSpace(void)
|
|
{
|
|
/* no large address space on win9x */
|
|
if (NtCurrentTeb()->Peb->OSPlatformId != VER_PLATFORM_WIN32_NT) return;
|
|
user_space_limit = ADDRESS_SPACE_LIMIT;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtAllocateVirtualMemory (NTDLL.@)
|
|
* ZwAllocateVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtAllocateVirtualMemory( HANDLE process, PVOID *ret, ULONG zero_bits,
|
|
SIZE_T *size_ptr, ULONG type, ULONG protect )
|
|
{
|
|
void *base;
|
|
BYTE vprot;
|
|
SIZE_T size = *size_ptr;
|
|
SIZE_T mask = get_mask( zero_bits );
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
struct file_view *view;
|
|
sigset_t sigset;
|
|
|
|
TRACE("%p %p %08lx %x %08x\n", process, *ret, size, type, protect );
|
|
|
|
if (!size) return STATUS_INVALID_PARAMETER;
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_alloc.type = APC_VIRTUAL_ALLOC;
|
|
call.virtual_alloc.addr = *ret;
|
|
call.virtual_alloc.size = *size_ptr;
|
|
call.virtual_alloc.zero_bits = zero_bits;
|
|
call.virtual_alloc.op_type = type;
|
|
call.virtual_alloc.prot = protect;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_alloc.status == STATUS_SUCCESS)
|
|
{
|
|
*ret = result.virtual_alloc.addr;
|
|
*size_ptr = result.virtual_alloc.size;
|
|
}
|
|
return result.virtual_alloc.status;
|
|
}
|
|
|
|
/* Round parameters to a page boundary */
|
|
|
|
if (size > 0x7fc00000) return STATUS_WORKING_SET_LIMIT_RANGE; /* 2Gb - 4Mb */
|
|
|
|
if (*ret)
|
|
{
|
|
if (type & MEM_RESERVE) /* Round down to 64k boundary */
|
|
base = ROUND_ADDR( *ret, mask );
|
|
else
|
|
base = ROUND_ADDR( *ret, page_mask );
|
|
size = (((UINT_PTR)*ret + size + page_mask) & ~page_mask) - (UINT_PTR)base;
|
|
|
|
/* disallow low 64k, wrap-around and kernel space */
|
|
if (((char *)base < (char *)0x10000) ||
|
|
((char *)base + size < (char *)base) ||
|
|
is_beyond_limit( base, size, ADDRESS_SPACE_LIMIT ))
|
|
return STATUS_INVALID_PARAMETER;
|
|
}
|
|
else
|
|
{
|
|
base = NULL;
|
|
size = (size + page_mask) & ~page_mask;
|
|
}
|
|
|
|
/* Compute the alloc type flags */
|
|
|
|
if (!(type & MEM_SYSTEM))
|
|
{
|
|
if (!(type & (MEM_COMMIT | MEM_RESERVE)) ||
|
|
(type & ~(MEM_COMMIT | MEM_RESERVE | MEM_TOP_DOWN | MEM_WRITE_WATCH | MEM_RESET)))
|
|
{
|
|
WARN("called with wrong alloc type flags (%08x) !\n", type);
|
|
return STATUS_INVALID_PARAMETER;
|
|
}
|
|
if (type & MEM_WRITE_WATCH)
|
|
{
|
|
FIXME("MEM_WRITE_WATCH type not supported\n");
|
|
return STATUS_NOT_SUPPORTED;
|
|
}
|
|
}
|
|
vprot = VIRTUAL_GetProt( protect );
|
|
if (type & MEM_COMMIT) vprot |= VPROT_COMMITTED;
|
|
|
|
/* Reserve the memory */
|
|
|
|
if (use_locks) server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (type & MEM_SYSTEM)
|
|
{
|
|
if (type & MEM_IMAGE) vprot |= VPROT_IMAGE;
|
|
status = create_view( &view, base, size, vprot | VPROT_COMMITTED );
|
|
if (status == STATUS_SUCCESS)
|
|
{
|
|
view->flags |= VFLAG_VALLOC | VFLAG_SYSTEM;
|
|
base = view->base;
|
|
}
|
|
}
|
|
else if ((type & MEM_RESERVE) || !base)
|
|
{
|
|
status = map_view( &view, base, size, mask, type & MEM_TOP_DOWN, vprot );
|
|
if (status == STATUS_SUCCESS)
|
|
{
|
|
view->flags |= VFLAG_VALLOC;
|
|
base = view->base;
|
|
}
|
|
}
|
|
else /* commit the pages */
|
|
{
|
|
if (!(view = VIRTUAL_FindView( base )) ||
|
|
((char *)base + size > (char *)view->base + view->size)) status = STATUS_NOT_MAPPED_VIEW;
|
|
else if (!VIRTUAL_SetProt( view, base, size, vprot )) status = STATUS_ACCESS_DENIED;
|
|
}
|
|
|
|
if (use_locks) server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (status == STATUS_SUCCESS)
|
|
{
|
|
*ret = base;
|
|
*size_ptr = size;
|
|
}
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtFreeVirtualMemory (NTDLL.@)
|
|
* ZwFreeVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtFreeVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr, ULONG type )
|
|
{
|
|
FILE_VIEW *view;
|
|
char *base;
|
|
sigset_t sigset;
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
LPVOID addr = *addr_ptr;
|
|
SIZE_T size = *size_ptr;
|
|
|
|
TRACE("%p %p %08lx %x\n", process, addr, size, type );
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_free.type = APC_VIRTUAL_FREE;
|
|
call.virtual_free.addr = addr;
|
|
call.virtual_free.size = size;
|
|
call.virtual_free.op_type = type;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_free.status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = result.virtual_free.addr;
|
|
*size_ptr = result.virtual_free.size;
|
|
}
|
|
return result.virtual_free.status;
|
|
}
|
|
|
|
/* Fix the parameters */
|
|
|
|
size = ROUND_SIZE( addr, size );
|
|
base = ROUND_ADDR( addr, page_mask );
|
|
|
|
/* avoid freeing the DOS area when a broken app passes a NULL pointer */
|
|
if (!base && !(type & MEM_SYSTEM)) return STATUS_INVALID_PARAMETER;
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (!(view = VIRTUAL_FindView( base )) ||
|
|
(base + size > (char *)view->base + view->size) ||
|
|
!(view->flags & VFLAG_VALLOC))
|
|
{
|
|
status = STATUS_INVALID_PARAMETER;
|
|
}
|
|
else if (type & MEM_SYSTEM)
|
|
{
|
|
/* return the values that the caller should use to unmap the area */
|
|
*addr_ptr = view->base;
|
|
if (!wine_mmap_is_in_reserved_area( view->base, view->size )) *size_ptr = view->size;
|
|
else *size_ptr = 0; /* make sure we don't munmap anything from a reserved area */
|
|
view->flags |= VFLAG_SYSTEM;
|
|
delete_view( view );
|
|
}
|
|
else if (type == MEM_RELEASE)
|
|
{
|
|
/* Free the pages */
|
|
|
|
if (size || (base != view->base)) status = STATUS_INVALID_PARAMETER;
|
|
else
|
|
{
|
|
delete_view( view );
|
|
*addr_ptr = base;
|
|
*size_ptr = size;
|
|
}
|
|
}
|
|
else if (type == MEM_DECOMMIT)
|
|
{
|
|
status = decommit_pages( view, base - (char *)view->base, size );
|
|
if (status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = base;
|
|
*size_ptr = size;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
WARN("called with wrong free type flags (%08x) !\n", type);
|
|
status = STATUS_INVALID_PARAMETER;
|
|
}
|
|
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtProtectVirtualMemory (NTDLL.@)
|
|
* ZwProtectVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtProtectVirtualMemory( HANDLE process, PVOID *addr_ptr, SIZE_T *size_ptr,
|
|
ULONG new_prot, ULONG *old_prot )
|
|
{
|
|
FILE_VIEW *view;
|
|
sigset_t sigset;
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
char *base;
|
|
UINT i;
|
|
BYTE vprot, *p;
|
|
ULONG prot;
|
|
SIZE_T size = *size_ptr;
|
|
LPVOID addr = *addr_ptr;
|
|
|
|
TRACE("%p %p %08lx %08x\n", process, addr, size, new_prot );
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_protect.type = APC_VIRTUAL_PROTECT;
|
|
call.virtual_protect.addr = addr;
|
|
call.virtual_protect.size = size;
|
|
call.virtual_protect.prot = new_prot;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_protect.status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = result.virtual_protect.addr;
|
|
*size_ptr = result.virtual_protect.size;
|
|
if (old_prot) *old_prot = result.virtual_protect.prot;
|
|
}
|
|
return result.virtual_protect.status;
|
|
}
|
|
|
|
/* Fix the parameters */
|
|
|
|
size = ROUND_SIZE( addr, size );
|
|
base = ROUND_ADDR( addr, page_mask );
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (!(view = VIRTUAL_FindView( base )) || (base + size > (char *)view->base + view->size))
|
|
{
|
|
status = STATUS_INVALID_PARAMETER;
|
|
}
|
|
else
|
|
{
|
|
/* Make sure all the pages are committed */
|
|
|
|
p = view->prot + ((base - (char *)view->base) >> page_shift);
|
|
prot = VIRTUAL_GetWin32Prot( *p );
|
|
for (i = size >> page_shift; i; i--, p++)
|
|
{
|
|
if (!(*p & VPROT_COMMITTED))
|
|
{
|
|
status = STATUS_NOT_COMMITTED;
|
|
break;
|
|
}
|
|
}
|
|
if (!i)
|
|
{
|
|
if (old_prot) *old_prot = prot;
|
|
vprot = VIRTUAL_GetProt( new_prot ) | VPROT_COMMITTED;
|
|
if (!VIRTUAL_SetProt( view, base, size, vprot )) status = STATUS_ACCESS_DENIED;
|
|
}
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
if (status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = base;
|
|
*size_ptr = size;
|
|
}
|
|
return status;
|
|
}
|
|
|
|
#define UNIMPLEMENTED_INFO_CLASS(c) \
|
|
case c: \
|
|
FIXME("(process=%p,addr=%p) Unimplemented information class: " #c "\n", process, addr); \
|
|
return STATUS_INVALID_INFO_CLASS
|
|
|
|
/***********************************************************************
|
|
* NtQueryVirtualMemory (NTDLL.@)
|
|
* ZwQueryVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtQueryVirtualMemory( HANDLE process, LPCVOID addr,
|
|
MEMORY_INFORMATION_CLASS info_class, PVOID buffer,
|
|
SIZE_T len, SIZE_T *res_len )
|
|
{
|
|
FILE_VIEW *view;
|
|
char *base, *alloc_base = 0;
|
|
struct list *ptr;
|
|
SIZE_T size = 0;
|
|
MEMORY_BASIC_INFORMATION *info = buffer;
|
|
sigset_t sigset;
|
|
|
|
if (info_class != MemoryBasicInformation)
|
|
{
|
|
switch(info_class)
|
|
{
|
|
UNIMPLEMENTED_INFO_CLASS(MemoryWorkingSetList);
|
|
UNIMPLEMENTED_INFO_CLASS(MemorySectionName);
|
|
UNIMPLEMENTED_INFO_CLASS(MemoryBasicVlmInformation);
|
|
|
|
default:
|
|
FIXME("(%p,%p,info_class=%d,%p,%ld,%p) Unknown information class\n",
|
|
process, addr, info_class, buffer, len, res_len);
|
|
return STATUS_INVALID_INFO_CLASS;
|
|
}
|
|
}
|
|
if (ADDRESS_SPACE_LIMIT && addr >= ADDRESS_SPACE_LIMIT)
|
|
return STATUS_WORKING_SET_LIMIT_RANGE;
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
NTSTATUS status;
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_query.type = APC_VIRTUAL_QUERY;
|
|
call.virtual_query.addr = addr;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_query.status == STATUS_SUCCESS)
|
|
{
|
|
info->BaseAddress = result.virtual_query.base;
|
|
info->AllocationBase = result.virtual_query.alloc_base;
|
|
info->RegionSize = result.virtual_query.size;
|
|
info->State = result.virtual_query.state;
|
|
info->Protect = result.virtual_query.prot;
|
|
info->AllocationProtect = result.virtual_query.alloc_prot;
|
|
info->Type = result.virtual_query.alloc_type;
|
|
if (res_len) *res_len = sizeof(*info);
|
|
}
|
|
return result.virtual_query.status;
|
|
}
|
|
|
|
base = ROUND_ADDR( addr, page_mask );
|
|
|
|
/* Find the view containing the address */
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
ptr = list_head( &views_list );
|
|
for (;;)
|
|
{
|
|
if (!ptr)
|
|
{
|
|
/* make the address space end at the user limit, except if
|
|
* the last view was mapped beyond that */
|
|
if (alloc_base <= (char *)user_space_limit)
|
|
{
|
|
if (user_space_limit && base >= (char *)user_space_limit)
|
|
{
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return STATUS_WORKING_SET_LIMIT_RANGE;
|
|
}
|
|
size = (char *)user_space_limit - alloc_base;
|
|
}
|
|
else size = (char *)ADDRESS_SPACE_LIMIT - alloc_base;
|
|
view = NULL;
|
|
break;
|
|
}
|
|
view = LIST_ENTRY( ptr, struct file_view, entry );
|
|
if ((char *)view->base > base)
|
|
{
|
|
size = (char *)view->base - alloc_base;
|
|
view = NULL;
|
|
break;
|
|
}
|
|
if ((char *)view->base + view->size > base)
|
|
{
|
|
alloc_base = view->base;
|
|
size = view->size;
|
|
break;
|
|
}
|
|
alloc_base = (char *)view->base + view->size;
|
|
ptr = list_next( &views_list, ptr );
|
|
}
|
|
|
|
/* Fill the info structure */
|
|
|
|
if (!view)
|
|
{
|
|
info->State = MEM_FREE;
|
|
info->Protect = PAGE_NOACCESS;
|
|
info->AllocationBase = 0;
|
|
info->AllocationProtect = 0;
|
|
info->Type = 0;
|
|
}
|
|
else
|
|
{
|
|
BYTE vprot = view->prot[(base - alloc_base) >> page_shift];
|
|
info->State = (vprot & VPROT_COMMITTED) ? MEM_COMMIT : MEM_RESERVE;
|
|
info->Protect = VIRTUAL_GetWin32Prot( vprot );
|
|
info->AllocationBase = alloc_base;
|
|
info->AllocationProtect = VIRTUAL_GetWin32Prot( view->protect );
|
|
if (view->protect & VPROT_IMAGE) info->Type = MEM_IMAGE;
|
|
else if (view->flags & VFLAG_VALLOC) info->Type = MEM_PRIVATE;
|
|
else info->Type = MEM_MAPPED;
|
|
for (size = base - alloc_base; size < view->size; size += page_size)
|
|
if (view->prot[size >> page_shift] != vprot) break;
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
info->BaseAddress = base;
|
|
info->RegionSize = size - (base - alloc_base);
|
|
if (res_len) *res_len = sizeof(*info);
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtLockVirtualMemory (NTDLL.@)
|
|
* ZwLockVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtLockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown )
|
|
{
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_lock.type = APC_VIRTUAL_LOCK;
|
|
call.virtual_lock.addr = *addr;
|
|
call.virtual_lock.size = *size;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_lock.status == STATUS_SUCCESS)
|
|
{
|
|
*addr = result.virtual_lock.addr;
|
|
*size = result.virtual_lock.size;
|
|
}
|
|
return result.virtual_lock.status;
|
|
}
|
|
|
|
*size = ROUND_SIZE( *addr, *size );
|
|
*addr = ROUND_ADDR( *addr, page_mask );
|
|
|
|
if (mlock( *addr, *size )) status = STATUS_ACCESS_DENIED;
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtUnlockVirtualMemory (NTDLL.@)
|
|
* ZwUnlockVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtUnlockVirtualMemory( HANDLE process, PVOID *addr, SIZE_T *size, ULONG unknown )
|
|
{
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_unlock.type = APC_VIRTUAL_UNLOCK;
|
|
call.virtual_unlock.addr = *addr;
|
|
call.virtual_unlock.size = *size;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_unlock.status == STATUS_SUCCESS)
|
|
{
|
|
*addr = result.virtual_unlock.addr;
|
|
*size = result.virtual_unlock.size;
|
|
}
|
|
return result.virtual_unlock.status;
|
|
}
|
|
|
|
*size = ROUND_SIZE( *addr, *size );
|
|
*addr = ROUND_ADDR( *addr, page_mask );
|
|
|
|
if (munlock( *addr, *size )) status = STATUS_ACCESS_DENIED;
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtCreateSection (NTDLL.@)
|
|
* ZwCreateSection (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtCreateSection( HANDLE *handle, ACCESS_MASK access, const OBJECT_ATTRIBUTES *attr,
|
|
const LARGE_INTEGER *size, ULONG protect,
|
|
ULONG sec_flags, HANDLE file )
|
|
{
|
|
NTSTATUS ret;
|
|
BYTE vprot;
|
|
DWORD len = (attr && attr->ObjectName) ? attr->ObjectName->Length : 0;
|
|
struct security_descriptor *sd = NULL;
|
|
struct object_attributes objattr;
|
|
|
|
/* Check parameters */
|
|
|
|
if (len > MAX_PATH*sizeof(WCHAR)) return STATUS_NAME_TOO_LONG;
|
|
|
|
objattr.rootdir = attr ? attr->RootDirectory : 0;
|
|
objattr.sd_len = 0;
|
|
objattr.name_len = len;
|
|
if (attr)
|
|
{
|
|
ret = NTDLL_create_struct_sd( attr->SecurityDescriptor, &sd, &objattr.sd_len );
|
|
if (ret != STATUS_SUCCESS) return ret;
|
|
}
|
|
|
|
vprot = VIRTUAL_GetProt( protect );
|
|
if (sec_flags & SEC_RESERVE)
|
|
{
|
|
if (file) return STATUS_INVALID_PARAMETER;
|
|
}
|
|
else vprot |= VPROT_COMMITTED;
|
|
if (sec_flags & SEC_NOCACHE) vprot |= VPROT_NOCACHE;
|
|
if (sec_flags & SEC_IMAGE) vprot |= VPROT_IMAGE;
|
|
|
|
/* Create the server object */
|
|
|
|
SERVER_START_REQ( create_mapping )
|
|
{
|
|
req->access = access;
|
|
req->attributes = (attr) ? attr->Attributes : 0;
|
|
req->file_handle = file;
|
|
req->size = size ? size->QuadPart : 0;
|
|
req->protect = vprot;
|
|
wine_server_add_data( req, &objattr, sizeof(objattr) );
|
|
if (objattr.sd_len) wine_server_add_data( req, sd, objattr.sd_len );
|
|
if (len) wine_server_add_data( req, attr->ObjectName->Buffer, len );
|
|
ret = wine_server_call( req );
|
|
*handle = reply->handle;
|
|
}
|
|
SERVER_END_REQ;
|
|
|
|
NTDLL_free_struct_sd( sd );
|
|
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtOpenSection (NTDLL.@)
|
|
* ZwOpenSection (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtOpenSection( HANDLE *handle, ACCESS_MASK access, const OBJECT_ATTRIBUTES *attr )
|
|
{
|
|
NTSTATUS ret;
|
|
DWORD len = attr->ObjectName->Length;
|
|
|
|
if (len > MAX_PATH*sizeof(WCHAR)) return STATUS_NAME_TOO_LONG;
|
|
|
|
SERVER_START_REQ( open_mapping )
|
|
{
|
|
req->access = access;
|
|
req->attributes = attr->Attributes;
|
|
req->rootdir = attr->RootDirectory;
|
|
wine_server_add_data( req, attr->ObjectName->Buffer, len );
|
|
if (!(ret = wine_server_call( req ))) *handle = reply->handle;
|
|
}
|
|
SERVER_END_REQ;
|
|
return ret;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtMapViewOfSection (NTDLL.@)
|
|
* ZwMapViewOfSection (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtMapViewOfSection( HANDLE handle, HANDLE process, PVOID *addr_ptr, ULONG zero_bits,
|
|
SIZE_T commit_size, const LARGE_INTEGER *offset_ptr, SIZE_T *size_ptr,
|
|
SECTION_INHERIT inherit, ULONG alloc_type, ULONG protect )
|
|
{
|
|
NTSTATUS res;
|
|
ULONGLONG full_size;
|
|
SIZE_T size = 0;
|
|
SIZE_T mask = get_mask( zero_bits );
|
|
int unix_handle = -1, needs_close;
|
|
int prot;
|
|
void *base;
|
|
struct file_view *view;
|
|
DWORD header_size;
|
|
HANDLE dup_mapping, shared_file;
|
|
LARGE_INTEGER offset;
|
|
sigset_t sigset;
|
|
|
|
offset.QuadPart = offset_ptr ? offset_ptr->QuadPart : 0;
|
|
|
|
TRACE("handle=%p process=%p addr=%p off=%x%08x size=%lx access=%x\n",
|
|
handle, process, *addr_ptr, offset.u.HighPart, offset.u.LowPart, size, protect );
|
|
|
|
/* Check parameters */
|
|
|
|
if ((offset.u.LowPart & mask) || (*addr_ptr && ((UINT_PTR)*addr_ptr & mask)))
|
|
return STATUS_INVALID_PARAMETER;
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.map_view.type = APC_MAP_VIEW;
|
|
call.map_view.handle = handle;
|
|
call.map_view.addr = *addr_ptr;
|
|
call.map_view.size = *size_ptr;
|
|
call.map_view.offset = offset.QuadPart;
|
|
call.map_view.zero_bits = zero_bits;
|
|
call.map_view.alloc_type = alloc_type;
|
|
call.map_view.prot = protect;
|
|
res = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (res != STATUS_SUCCESS) return res;
|
|
|
|
if (result.map_view.status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = result.map_view.addr;
|
|
*size_ptr = result.map_view.size;
|
|
}
|
|
return result.map_view.status;
|
|
}
|
|
|
|
SERVER_START_REQ( get_mapping_info )
|
|
{
|
|
req->handle = handle;
|
|
res = wine_server_call( req );
|
|
prot = reply->protect;
|
|
base = reply->base;
|
|
full_size = reply->size;
|
|
header_size = reply->header_size;
|
|
dup_mapping = reply->mapping;
|
|
shared_file = reply->shared_file;
|
|
}
|
|
SERVER_END_REQ;
|
|
if (res) return res;
|
|
|
|
size = full_size;
|
|
if (sizeof(size) < sizeof(full_size) && (size != full_size))
|
|
ERR( "Sizes larger than 4Gb (%x%08x) not supported on this platform\n",
|
|
(DWORD)(full_size >> 32), (DWORD)full_size );
|
|
|
|
if ((res = server_get_unix_fd( handle, 0, &unix_handle, &needs_close, NULL, NULL ))) goto done;
|
|
|
|
if (prot & VPROT_IMAGE)
|
|
{
|
|
if (shared_file)
|
|
{
|
|
int shared_fd, shared_needs_close;
|
|
|
|
if ((res = server_get_unix_fd( shared_file, FILE_READ_DATA|FILE_WRITE_DATA,
|
|
&shared_fd, &shared_needs_close, NULL, NULL ))) goto done;
|
|
res = map_image( handle, unix_handle, base, size, mask, header_size,
|
|
shared_fd, dup_mapping, addr_ptr );
|
|
if (shared_needs_close) close( shared_fd );
|
|
NtClose( shared_file );
|
|
}
|
|
else
|
|
{
|
|
res = map_image( handle, unix_handle, base, size, mask, header_size,
|
|
-1, dup_mapping, addr_ptr );
|
|
}
|
|
if (needs_close) close( unix_handle );
|
|
if (!res) *size_ptr = size;
|
|
return res;
|
|
}
|
|
|
|
if ((offset.QuadPart >= size) || (*size_ptr > size - offset.QuadPart))
|
|
{
|
|
res = STATUS_INVALID_PARAMETER;
|
|
goto done;
|
|
}
|
|
if (*size_ptr) size = ROUND_SIZE( offset.u.LowPart, *size_ptr );
|
|
else size = size - offset.QuadPart;
|
|
|
|
switch(protect)
|
|
{
|
|
case PAGE_NOACCESS:
|
|
break;
|
|
case PAGE_READWRITE:
|
|
case PAGE_EXECUTE_READWRITE:
|
|
if (!(prot & VPROT_WRITE))
|
|
{
|
|
res = STATUS_INVALID_PARAMETER;
|
|
goto done;
|
|
}
|
|
/* fall through */
|
|
case PAGE_READONLY:
|
|
case PAGE_WRITECOPY:
|
|
case PAGE_EXECUTE:
|
|
case PAGE_EXECUTE_READ:
|
|
case PAGE_EXECUTE_WRITECOPY:
|
|
if (prot & VPROT_READ) break;
|
|
/* fall through */
|
|
default:
|
|
res = STATUS_INVALID_PARAMETER;
|
|
goto done;
|
|
}
|
|
|
|
/* FIXME: If a mapping is created with SEC_RESERVE and a process,
|
|
* which has a view of this mapping commits some pages, they will
|
|
* appear committed in all other processes, which have the same
|
|
* view created. Since we don't support this yet, we create the
|
|
* whole mapping committed.
|
|
*/
|
|
prot |= VPROT_COMMITTED;
|
|
|
|
/* Reserve a properly aligned area */
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
res = map_view( &view, *addr_ptr, size, mask, FALSE, prot );
|
|
if (res)
|
|
{
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
goto done;
|
|
}
|
|
|
|
/* Map the file */
|
|
|
|
TRACE("handle=%p size=%lx offset=%x%08x\n",
|
|
handle, size, offset.u.HighPart, offset.u.LowPart );
|
|
|
|
res = map_file_into_view( view, unix_handle, 0, size, offset.QuadPart, prot, !dup_mapping );
|
|
if (res == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = view->base;
|
|
*size_ptr = size;
|
|
view->mapping = dup_mapping;
|
|
dup_mapping = 0; /* don't close it */
|
|
}
|
|
else
|
|
{
|
|
ERR( "map_file_into_view %p %lx %x%08x failed\n",
|
|
view->base, size, offset.u.HighPart, offset.u.LowPart );
|
|
delete_view( view );
|
|
}
|
|
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
|
|
done:
|
|
if (dup_mapping) NtClose( dup_mapping );
|
|
if (needs_close) close( unix_handle );
|
|
return res;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtUnmapViewOfSection (NTDLL.@)
|
|
* ZwUnmapViewOfSection (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtUnmapViewOfSection( HANDLE process, PVOID addr )
|
|
{
|
|
FILE_VIEW *view;
|
|
NTSTATUS status = STATUS_INVALID_PARAMETER;
|
|
sigset_t sigset;
|
|
void *base = ROUND_ADDR( addr, page_mask );
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.unmap_view.type = APC_UNMAP_VIEW;
|
|
call.unmap_view.addr = addr;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status == STATUS_SUCCESS) status = result.unmap_view.status;
|
|
return status;
|
|
}
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
if ((view = VIRTUAL_FindView( base )) && (base == view->base))
|
|
{
|
|
delete_view( view );
|
|
status = STATUS_SUCCESS;
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtFlushVirtualMemory (NTDLL.@)
|
|
* ZwFlushVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtFlushVirtualMemory( HANDLE process, LPCVOID *addr_ptr,
|
|
SIZE_T *size_ptr, ULONG unknown )
|
|
{
|
|
FILE_VIEW *view;
|
|
NTSTATUS status = STATUS_SUCCESS;
|
|
sigset_t sigset;
|
|
void *addr = ROUND_ADDR( *addr_ptr, page_mask );
|
|
|
|
if (process != NtCurrentProcess())
|
|
{
|
|
apc_call_t call;
|
|
apc_result_t result;
|
|
|
|
memset( &call, 0, sizeof(call) );
|
|
|
|
call.virtual_flush.type = APC_VIRTUAL_FLUSH;
|
|
call.virtual_flush.addr = addr;
|
|
call.virtual_flush.size = *size_ptr;
|
|
status = NTDLL_queue_process_apc( process, &call, &result );
|
|
if (status != STATUS_SUCCESS) return status;
|
|
|
|
if (result.virtual_flush.status == STATUS_SUCCESS)
|
|
{
|
|
*addr_ptr = result.virtual_flush.addr;
|
|
*size_ptr = result.virtual_flush.size;
|
|
}
|
|
return result.virtual_flush.status;
|
|
}
|
|
|
|
server_enter_uninterrupted_section( &csVirtual, &sigset );
|
|
if (!(view = VIRTUAL_FindView( addr ))) status = STATUS_INVALID_PARAMETER;
|
|
else
|
|
{
|
|
if (!*size_ptr) *size_ptr = view->size;
|
|
*addr_ptr = addr;
|
|
if (msync( addr, *size_ptr, MS_SYNC )) status = STATUS_NOT_MAPPED_DATA;
|
|
}
|
|
server_leave_uninterrupted_section( &csVirtual, &sigset );
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtReadVirtualMemory (NTDLL.@)
|
|
* ZwReadVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtReadVirtualMemory( HANDLE process, const void *addr, void *buffer,
|
|
SIZE_T size, SIZE_T *bytes_read )
|
|
{
|
|
NTSTATUS status;
|
|
|
|
SERVER_START_REQ( read_process_memory )
|
|
{
|
|
req->handle = process;
|
|
req->addr = (void *)addr;
|
|
wine_server_set_reply( req, buffer, size );
|
|
if ((status = wine_server_call( req ))) size = 0;
|
|
}
|
|
SERVER_END_REQ;
|
|
if (bytes_read) *bytes_read = size;
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtWriteVirtualMemory (NTDLL.@)
|
|
* ZwWriteVirtualMemory (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtWriteVirtualMemory( HANDLE process, void *addr, const void *buffer,
|
|
SIZE_T size, SIZE_T *bytes_written )
|
|
{
|
|
NTSTATUS status;
|
|
|
|
SERVER_START_REQ( write_process_memory )
|
|
{
|
|
req->handle = process;
|
|
req->addr = addr;
|
|
wine_server_add_data( req, buffer, size );
|
|
if ((status = wine_server_call( req ))) size = 0;
|
|
}
|
|
SERVER_END_REQ;
|
|
if (bytes_written) *bytes_written = size;
|
|
return status;
|
|
}
|
|
|
|
|
|
/***********************************************************************
|
|
* NtAreMappedFilesTheSame (NTDLL.@)
|
|
* ZwAreMappedFilesTheSame (NTDLL.@)
|
|
*/
|
|
NTSTATUS WINAPI NtAreMappedFilesTheSame(PVOID addr1, PVOID addr2)
|
|
{
|
|
TRACE("%p %p\n", addr1, addr2);
|
|
|
|
return STATUS_NOT_SAME_DEVICE;
|
|
}
|