blob: 330b674ea73bbb585a15c454c1b3a6f2a66b26d4 [file] [log] [blame]
/* **********************************************************
* Copyright (c) 2010-2013 Google, Inc. All rights reserved.
* Copyright (c) 2001-2010 VMware, Inc. All rights reserved.
* **********************************************************/
/*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions are met:
*
* * Redistributions of source code must retain the above copyright notice,
* this list of conditions and the following disclaimer.
*
* * Redistributions in binary form must reproduce the above copyright notice,
* this list of conditions and the following disclaimer in the documentation
* and/or other materials provided with the distribution.
*
* * Neither the name of VMware, Inc. nor the names of its contributors may be
* used to endorse or promote products derived from this software without
* specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL VMWARE, INC. OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
* SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH
* DAMAGE.
*/
/* Copyright (c) 2003-2007 Determina Corp. */
/* Copyright (c) 2001-2003 Massachusetts Institute of Technology */
/* Copyright (c) 2001 Hewlett-Packard Company */
/*
* heap.c - heap manager
*/
#include "globals.h"
#include <string.h> /* for memcpy */
#include <limits.h>
#include "fragment.h" /* for struct sizes */
#include "link.h" /* for struct sizes */
#include "instr.h" /* for struct sizes */
#include "fcache.h" /* fcache_low_on_memory */
#ifdef DEBUG
# include "hotpatch.h" /* To handle leak for case 9593. */
#endif
#ifdef HEAP_ACCOUNTING
# ifndef DEBUG
# error HEAP_ACCOUNTING requires DEBUG
# endif
#endif
#ifdef DEBUG_MEMORY
/* on by default but higher than general asserts */
# define CHKLVL_MEMFILL CHKLVL_DEFAULT
#endif
extern bool vm_areas_exited;
/***************************************************************************
* we cannot use malloc in the middle of interpreting the client program
* because we could be in the middle of interpreting malloc, which is not
* always reentrant
*
* We have a virtual memory manager which makes sure memory is
* reserved within the application address space so that we don't have
* to fight with the application. We call os_heap_reserve to allocate
* virtual space in a single consecutive region. We later use
* os_heap_commit to get committed memory in large chunks and manage
* the chunks using a simple scheme of free lists of different sizes.
* The virtual memory manager has to store out of band information about
* used and free blocks, since of course there is no real memory to use.
* The chunks (heap units) store in band extra information both for
* used and free. However, in the allocated blocks within a unit we
* don't need to store any information since heap_free passes in the
* size; we store the next pointers for the free lists at the start of
* the free blocks themselves. We have one large reservation for most of
* our allocations, and yet another for allocations that we do not
* plan on ever freeing up on detach - the only unavoidable tombstones
* are those for thread private code system calls that may be stuck on
* callbacks. In case we run out of reserved memory we do fall back
* on requests from the OS, but any of these may fail if we are
* competing with the application.
*
* looking at dynamo behavior as of Jan 2001, most heap_alloc requests are
* for < 128 bytes, very few for larger, so we have a bunch of fixed-size
* blocks of small sizes
*
* the UINT_MAX size is a variable-length block, we keep one byte to store
* the size (again storing the next pointer when free at the start of
* what we pass to the user)
*/
static const uint BLOCK_SIZES[] = {
8, /* for instr bits */
#ifndef X64
/* for x64 future_fragment_t is 24 bytes (could be 20 if we could put flags last) */
sizeof(future_fragment_t), /* 12 (24 x64) */
#endif
/* we have a lot of size 16 requests for IR but they are transient */
24, /* fcache empties and vm_area_t are now 20, vm area extras still 24 */
ALIGN_FORWARD(sizeof(fragment_t) + sizeof(indirect_linkstub_t), HEAP_ALIGNMENT), /* 40 dbg / 36 rel */
#if defined(X64) || defined(PROFILE_LINKCOUNT) || defined(CUSTOM_EXIT_STUBS)
sizeof(instr_t), /* 64 (104 x64) */
sizeof(fragment_t) + sizeof(direct_linkstub_t)
+ sizeof(cbr_fallthrough_linkstub_t), /* 68 dbg / 64 rel, 112 x64 */
/* all other bb/trace buckets are 8 larger but in same order */
#else
sizeof(fragment_t) + sizeof(direct_linkstub_t)
+ sizeof(cbr_fallthrough_linkstub_t), /* 60 dbg / 56 rel */
sizeof(instr_t), /* 64 */
#endif
/* we keep this bucket even though only 10% or so of normal bbs
* hit this.
* FIXME: release == instr_t here so a small waste when walking buckets
*/
ALIGN_FORWARD(sizeof(fragment_t) + 2*sizeof(direct_linkstub_t),
HEAP_ALIGNMENT), /* 68 dbg / 64 rel (128 x64) */
ALIGN_FORWARD(sizeof(trace_t) + 2*sizeof(direct_linkstub_t) + sizeof(uint),
HEAP_ALIGNMENT), /* 80 dbg / 76 rel (148 x64 => 152) */
/* FIXME: measure whether should put in indirect mixes as well */
ALIGN_FORWARD(sizeof(trace_t) + 3*sizeof(direct_linkstub_t) + sizeof(uint),
HEAP_ALIGNMENT), /* 96 dbg / 92 rel (180 x64 => 184) */
ALIGN_FORWARD(sizeof(trace_t) + 5*sizeof(direct_linkstub_t) + sizeof(uint),
HEAP_ALIGNMENT), /* 128 dbg / 124 rel (244 x64 => 248) */
256,
512,
UINT_MAX /* variable-length */
};
#define BLOCK_TYPES (sizeof(BLOCK_SIZES)/sizeof(uint))
#ifdef DEBUG
/* FIXME: would be nice to have these stats per HEAPACCT category */
/* These are ints only b/c we used to do non-atomic adds and wanted to
* gracefully handle underflow to negative values
*/
DECLARE_NEVERPROT_VAR(static int block_total_count[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_count[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_peak_count[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_wasted[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_peak_wasted[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_align_pad[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static int block_peak_align_pad[BLOCK_TYPES], {0});
DECLARE_NEVERPROT_VAR(static bool out_of_vmheap_once, false);
#endif
/* variable-length: we steal one int for the size */
#define HEADER_SIZE (sizeof(size_t))
/* VARIABLE_SIZE is assignable */
#define VARIABLE_SIZE(p) (*(size_t *)((p)-HEADER_SIZE))
#define MEMSET_HEADER(p, value) VARIABLE_SIZE(p) = HEAP_TO_PTR_UINT(value)
#define GET_VARIABLE_ALLOCATION_SIZE(p) (VARIABLE_SIZE(p) + HEADER_SIZE)
/* heap is allocated in units
* we start out with a small unit, then each additional unit we
* need doubles in size, up to a maximum, we default to 32kb initial size
* (24kb useful with guard pages), max size defaults to 64kb (56kb useful with
* guard pages), we keep the max small to save memory, it doesn't seem to be
* perf hit! Though with guard pages we are wasting quite a bit of reserved
* (though not committed) space */
/* the only big things global heap is used for are pc sampling
* hash table and sideline sampling hash table -- if none of those
* are in use, 16KB should be plenty, we default to 32kb since guard
* pages are on by default (gives 24kb useful) max size is same as for
* normal heap units.
*/
/* the old defaults were 32kb (usable) for initial thread private and 16kb
* (usable) for initial global, changed to simplify the logic for allocating
* in multiples of the os allocation granularity. The new defaults prob.
* make more sense with the shared cache then the old did anyways.
*/
/* restrictions -
* any guard pages are included in the size, size must be > UNITOVERHEAD
* for best performance sizes should be of the form
* 2^n * page_size (where n is a positve integer) and max should be a multiple
* of the os allocation granularity so that with enough doublings we are
* reserving memory in multiples of the allocation granularity and not wasting
* any virtual address space (beyond our guard pages)
*/
#define HEAP_UNIT_MIN_SIZE INTERNAL_OPTION(initial_heap_unit_size)
#define HEAP_UNIT_MAX_SIZE INTERNAL_OPTION(max_heap_unit_size)
#define GLOBAL_UNIT_MIN_SIZE INTERNAL_OPTION(initial_global_heap_unit_size)
#define GUARD_PAGE_ADJUSTMENT (dynamo_options.guard_pages ? 2 * PAGE_SIZE : 0)
/* gets usable space in the unit */
#define UNITROOM(u) ((size_t) (u->end_pc - u->start_pc))
#define UNIT_RESERVED_ROOM(u) (u->reserved_end_pc - u->start_pc)
/* we keep the heap_unit_t header at top of the unit, this macro calculates
* the committed size of the unit by adding header size to available size
*/
#define UNIT_COMMIT_SIZE(u) (UNITROOM(u) + sizeof(heap_unit_t))
#define UNIT_RESERVED_SIZE(u) (UNIT_RESERVED_ROOM(u) + sizeof(heap_unit_t))
#define UNIT_ALLOC_START(u) (u->start_pc - sizeof(heap_unit_t))
#define UNIT_GET_START_PC(u) (byte*)(((ptr_uint_t)u) + sizeof(heap_unit_t))
#define UNIT_COMMIT_END(u) (u->end_pc)
#define UNIT_RESERVED_END(u) (u->reserved_end_pc)
/* gets the allocated size of the unit (reserved size + guard pages) */
#define UNITALLOC(u) (UNIT_RESERVED_SIZE(u) + GUARD_PAGE_ADJUSTMENT)
/* gets unit overhead, includes the reserved (guard pages) and committed
* (sizeof(heap_unit_t)) portions
*/
#define UNITOVERHEAD (sizeof(heap_unit_t) + GUARD_PAGE_ADJUSTMENT)
/* any alloc request larger than this needs a special unit */
#define MAXROOM (HEAP_UNIT_MAX_SIZE - UNITOVERHEAD)
/* maximum valid allocation (to guard against internal integer overflows) */
#define MAX_VALID_HEAP_ALLOCATION INT_MAX
/* thread-local heap structure
* this struct is kept at top of unit itself, not in separate allocation
*/
typedef struct _heap_unit_t {
heap_pc start_pc; /* start address of heap storage */
heap_pc end_pc; /* open-ended end address of heap storage */
heap_pc cur_pc; /* open-ended current end of allocated storage */
heap_pc reserved_end_pc; /* open-ended end of reserved (not nec committed) memory */
bool in_vmarea_list; /* perf opt for delayed batch vmarea updating */
#ifdef DEBUG
int id; /* # of this unit */
#endif
struct _heap_unit_t *next_local; /* used to link thread's units */
struct _heap_unit_t *next_global; /* used to link all units */
struct _heap_unit_t *prev_global; /* used to link all units */
} heap_unit_t;
#ifdef HEAP_ACCOUNTING
typedef struct _heap_acct_t {
size_t alloc_reuse[ACCT_LAST];
size_t alloc_new[ACCT_LAST];
size_t cur_usage[ACCT_LAST];
size_t max_usage[ACCT_LAST];
size_t max_single[ACCT_LAST];
uint num_alloc[ACCT_LAST];
} heap_acct_t;
#endif
/* FIXME (case 6336): rename to heap_t:
* a heap_t is a collection of units with the same properties
* to reflect that this is used for more than just thread-private memory.
* Also rename the "tu" vars to "h"
*/
typedef struct _thread_units_t {
heap_unit_t *top_unit; /* start of linked list of heap units */
heap_unit_t *cur_unit; /* current unit in heap list */
heap_pc free_list[BLOCK_TYPES];
#ifdef DEBUG
int num_units; /* total # of heap units */
#endif
dcontext_t *dcontext; /* back pointer to owner */
bool writable; /* remember state of heap protection */
#ifdef HEAP_ACCOUNTING
heap_acct_t acct;
#endif
} thread_units_t;
/* per-thread structure: */
typedef struct _thread_heap_t {
thread_units_t *local_heap;
thread_units_t *nonpersistent_heap;
} thread_heap_t;
/* global, unique thread-shared structure:
* FIXME: give this name to thread_units_t, and name this AllHeapUnits
*/
typedef struct _heap_t {
heap_unit_t *units; /* list of all allocated units */
heap_unit_t *dead; /* list of deleted units ready for re-allocation */
/* FIXME: num_dead duplicates stats->heap_num_free, but we want num_dead
* for release build too, so it's separate...can we do better?
*/
uint num_dead;
} heap_t;
/* no synch needed since only written once */
static bool heap_exiting = false;
#ifdef DEBUG
DECLARE_NEVERPROT_VAR(static bool ever_beyond_vmm, false);
#endif
/* Lock used only for managing heap units, not for normal thread-local alloc.
* Must be recursive due to circular dependencies between vmareas and global heap.
* Furthermore, always grab dynamo_vm_areas_lock() before grabbing this lock,
* to make DR areas update and heap alloc/free atomic!
*/
DECLARE_CXTSWPROT_VAR(static recursive_lock_t heap_unit_lock,
INIT_RECURSIVE_LOCK(heap_unit_lock));
/* N.B.: if these two locks are ever owned at the same time, the convention is
* that global_alloc_lock MUST be grabbed first, to avoid deadlocks
*/
/* separate lock for global heap access to avoid contention between local unit
* creation and global heap alloc
* must be recursive so that heap_vmareas_synch_units can hold it and heap_unit_lock
* up front to avoid deadlocks, and still allow vmareas to global_alloc --
* BUT we do NOT want global_alloc() to be able to recurse!
* FIXME: either find a better solution to the heap_vmareas_synch_units deadlock
* that is as efficient, or find a way to assert that the only recursion is
* from heap_vmareas_synch_units to global_alloc
*/
DECLARE_CXTSWPROT_VAR(static recursive_lock_t global_alloc_lock,
INIT_RECURSIVE_LOCK(global_alloc_lock));
#if defined(DEBUG) && defined(HEAP_ACCOUNTING) && defined(HOT_PATCHING_INTERFACE)
static int get_special_heap_header_size(void);
#endif
vm_area_vector_t *landing_pad_areas; /* PR 250294 */
#ifdef WINDOWS
/* i#939: we steal space from ntdll's +rx segment */
static app_pc lpad_temp_writable_start;
static size_t lpad_temp_writable_size;
static void release_landing_pad_mem(void);
#endif
/* Indicates whether should back out of a global alloc/free and grab the
* DR areas lock first, to retry
*/
static bool
safe_to_allocate_or_free_heap_units()
{
return ((!self_owns_recursive_lock(&global_alloc_lock) &&
!self_owns_recursive_lock(&heap_unit_lock)) ||
self_owns_dynamo_vm_area_lock());
}
/* indicates a dynamo vm area remove was delayed
* protected by the heap_unit_lock
*/
DECLARE_FREQPROT_VAR(static bool dynamo_areas_pending_remove, false);
#ifdef HEAP_ACCOUNTING
const char * whichheap_name[] = {
/* max length for aligned output is length of "BB Fragments" */
"BB Fragments",
"Coarse Links",
"Future Frag",
"Frag Tables",
"IBL Tables",
"Traces",
"FC Empties",
"Vm Multis",
"IR",
"RCT Tables",
"VM Areas",
"Symbols",
# ifdef SIDELINE
"Sideline",
# endif
"TH Counter",
"Tombstone",
"Hot Patching",
"Thread Mgt",
"Memory Mgt",
"Stats",
"SpecialHeap",
# ifdef CLIENT_INTERFACE
"Client",
# endif
"Lib Dup",
"Clean Call",
/* NOTE: Add your heap name here */
"Other",
};
/* Since using a lock for these stats adds a lot of contention, we
* follow a two-pronged strategy:
* 1) For accurate stats we add a thread's final stats to the global only
* when it is cleaned up. But, this prevents global stats from being
* available in the middle of a run or if a run is not cleaned up nicely.
* 2) We have a set of heap_accounting stats for incremental global stats
* that are available at any time, yet racy and so may be off a little.
*/
/* all set to 0 is only initialization we need */
DECLARE_NEVERPROT_VAR(static thread_units_t global_racy_units, {0});
/* macro to get the type abstracted */
# define ACCOUNT_FOR_ALLOC_HELPER(type, tu, which, alloc_sz, ask_sz) do { \
(tu)->acct.type[which] += alloc_sz; \
(tu)->acct.num_alloc[which]++; \
(tu)->acct.cur_usage[which] += alloc_sz; \
if ((tu)->acct.cur_usage[which] > (tu)->acct.max_usage[which]) \
(tu)->acct.max_usage[which] = (tu)->acct.cur_usage[which]; \
if (ask_sz > (tu)->acct.max_single[which]) \
(tu)->acct.max_single[which] = ask_sz; \
} while (0)
# define ACCOUNT_FOR_ALLOC(type, tu, which, alloc_sz, ask_sz) do { \
STATS_ADD_PEAK(heap_claimed, alloc_sz); \
ACCOUNT_FOR_ALLOC_HELPER(type, tu, which, alloc_sz, ask_sz); \
ACCOUNT_FOR_ALLOC_HELPER(type, &global_racy_units, which, \
alloc_sz, ask_sz); \
} while (0)
# define ACCOUNT_FOR_FREE(tu, which, size) do { \
STATS_SUB(heap_claimed, (size)); \
(tu)->acct.cur_usage[which] -= size; \
global_racy_units.acct.cur_usage[which] -= size; \
} while (0)
#else
# define ACCOUNT_FOR_ALLOC(type, tu, which, alloc_sz, ask_sz)
# define ACCOUNT_FOR_FREE(tu, which, size)
#endif
typedef byte *vm_addr_t;
#ifdef X64
/* designates the closed interval within which we must allocate DR heap space */
static byte *heap_allowable_region_start = (byte *)PTR_UINT_0;
static byte *heap_allowable_region_end = (byte *)POINTER_MAX;
/* used only to protect read/write access to the must_reach_* static variables in
* request_region_be_heap_reachable() */
DECLARE_CXTSWPROT_VAR(static mutex_t request_region_be_heap_reachable_lock,
INIT_LOCK_FREE(request_region_be_heap_reachable_lock));
/* Request that the supplied region be 32bit offset reachable from the DR heap. Should
* be called before vmm_heap_init() so we can place the DR heap to meet these constraints.
* Can also be called post vmm_heap_init() but at that point acts as an assert that the
* supplied region is reachable since the heap is already reserved.
*
* Must be called at least once up front, for the -heap_in_lower_4GB code here
* to kick in!
*/
void
request_region_be_heap_reachable(byte *start, size_t size)
{
/* initialize so will be overridden on first call; protected by the
* request_region_be_heap_reachable_lock */
static byte *must_reach_region_start = (byte *)POINTER_MAX;
static byte *must_reach_region_end = (byte *)PTR_UINT_0; /* closed */
LOG(GLOBAL, LOG_HEAP, 2,
"Adding must-be-reachable-from-heap region "PFX"-"PFX"\n"
"Existing must-be-reachable region "PFX"-"PFX"\n"
"Existing allowed range "PFX"-"PFX"\n",
start, start+size, must_reach_region_start, must_reach_region_end,
heap_allowable_region_start, heap_allowable_region_end);
ASSERT(!POINTER_OVERFLOW_ON_ADD(start, size));
ASSERT(size > 0);
mutex_lock(&request_region_be_heap_reachable_lock);
if (start < must_reach_region_start) {
byte *allowable_end_tmp;
SELF_UNPROTECT_DATASEC(DATASEC_RARELY_PROT);
must_reach_region_start = start;
allowable_end_tmp = REACHABLE_32BIT_END(must_reach_region_start,
must_reach_region_end);
/* PR 215395 - add in absolute address reachability */
if (DYNAMO_OPTION(heap_in_lower_4GB) &&
allowable_end_tmp > ( byte *)POINTER_MAX_32BIT) {
allowable_end_tmp = (byte *)POINTER_MAX_32BIT;
}
/* Write assumed to be atomic so we don't have to hold a lock to use
* heap_allowable_region_end. */
heap_allowable_region_end = allowable_end_tmp;
SELF_PROTECT_DATASEC(DATASEC_RARELY_PROT);
}
if (start + size - 1 > must_reach_region_end) {
SELF_UNPROTECT_DATASEC(DATASEC_RARELY_PROT);
must_reach_region_end = start + size - 1; /* closed */
/* Write assumed to be atomic so we don't have to hold a lock to use
* heap_allowable_region_start. */
heap_allowable_region_start = REACHABLE_32BIT_START(must_reach_region_start,
must_reach_region_end);
SELF_PROTECT_DATASEC(DATASEC_RARELY_PROT);
}
ASSERT(must_reach_region_start <= must_reach_region_end); /* correctness check */
/* verify can be addressed absolutely (if required), correctness check */
ASSERT(!DYNAMO_OPTION(heap_in_lower_4GB) ||
heap_allowable_region_end <= (byte *)POINTER_MAX_32BIT);
mutex_unlock(&request_region_be_heap_reachable_lock);
LOG(GLOBAL, LOG_HEAP, 1,
"Added must-be-reachable-from-heap region "PFX"-"PFX"\n"
"New must-be-reachable region "PFX"-"PFX"\n"
"New allowed range "PFX"-"PFX"\n",
start, start+size, must_reach_region_start, must_reach_region_end,
heap_allowable_region_start, heap_allowable_region_end);
/* Reachability checks (xref PR 215395, note since we currently can't directly
* control where DR/client dlls are loaded these could fire if rebased). */
ASSERT(heap_allowable_region_start <= must_reach_region_start &&
"x64 reachability contraints not satisfiable");
ASSERT(must_reach_region_end <= heap_allowable_region_end &&
"x64 reachability contraints not satisfiable");
/* Handle release build failure. */
if (heap_allowable_region_start > must_reach_region_start ||
must_reach_region_end > heap_allowable_region_end) {
/* FIXME - in a released product we may want to detach or something else less
* drastic than triggering a FATAL_USAGE_ERROR. */
FATAL_USAGE_ERROR(HEAP_CONTRAINTS_UNSATISFIABLE, 2,
get_application_name(), get_application_pid());
}
}
void
vmcode_get_reachable_region(byte **region_start OUT, byte **region_end OUT)
{
/* We track sub-page for more accuracy on additional constraints, and
* align when asked about it.
*/
if (region_start != NULL)
*region_start = (byte *) ALIGN_FORWARD(heap_allowable_region_start, PAGE_SIZE);
if (region_end != NULL)
*region_end = (byte *) ALIGN_BACKWARD(heap_allowable_region_end, PAGE_SIZE);
}
#endif
/* forward declarations of static functions */
static void threadunits_init(dcontext_t *dcontext, thread_units_t *tu, size_t size);
/* dcontext only used for debugging */
static void threadunits_exit(thread_units_t *tu, dcontext_t *dcontext);
static void *common_heap_alloc(thread_units_t *tu, size_t size
HEAPACCT(which_heap_t which));
static bool common_heap_free(thread_units_t *tu, void *p, size_t size
HEAPACCT(which_heap_t which));
static void release_real_memory(void *p, size_t size, bool remove_vm);
static void release_guarded_real_memory(vm_addr_t p, size_t size, bool remove_vm,
bool guarded);
typedef enum {
/* I - Init, Interop - first allocation failed
* check for incompatible kernel drivers
*/
OOM_INIT = 0x1,
/* R - Reserve - out of virtual reservation *
* increase -vm_size to reserve more memory
*/
OOM_RESERVE = 0x2,
/* C - Commit - systemwide page file limit, or current process job limit hit
* Increase pagefile size, check for memory leak in any application.
*
* FIXME: possible automatic actions
* if systemwide failure we may want to wait if transient
* FIXME: if in a job latter we want to detect and just die
* (though after freeing as much memory as we can)
*/
OOM_COMMIT = 0x4,
/* E - Extending Commit - same reasons as Commit
* as a possible workaround increasing -heap_commit_increment
* may make expose us to commit-ing less frequently,
* On the other hand committing smaller chunks has a higher
* chance of getting through when there is very little memory.
*
* FIXME: not much more informative than OOM_COMMIT
*/
OOM_EXTEND = 0x8,
} oom_source_t;
static void report_low_on_memory(oom_source_t source,
heap_error_code_t os_error_code);
/* virtual memory manager */
enum {VMM_BLOCK_SIZE = IF_WINDOWS_ELSE(64,16)*1024}; /* 64KB or 16KB */
/* Our current allocation unit matches the allocation granularity on
* windows, to avoid worrying about external fragmentation
* Since most of our allocations fall within this range this makes the
* common operation be finding a single empty block.
*
* On Linux we save a lot of wasted alignment space by using a smaller
* granularity (PR 415959).
*
* FIXME: for Windows, if we reserve the whole region up front and
* just commit pieces, why do we need to match the Windows kernel
* alloc granularity?
*/
enum {
/* maximum 512MB virtual memory units */
MAX_VMM_HEAP_UNIT_SIZE = 512*1024*1024,
/* We should normally have only one large unit, so this is in fact
* the maximum we should count on in one process
*/
/* minimum will be used only if an invalid option is set */
MIN_VMM_HEAP_UNIT_SIZE = VMM_BLOCK_SIZE
};
typedef struct {
vm_addr_t start_addr; /* base virtual address */
vm_addr_t end_addr; /* noninclusive virtual memory range [start,end) */
vm_addr_t alloc_start; /* base allocation virtual address */
size_t alloc_size; /* allocation size */
/* for 64-bit do we want to shift to size_t to allow a larger region?
* if so must update the bitmap_t routines
*/
uint num_blocks; /* total number of blocks in virtual allocation */
mutex_t lock; /* write access to the rest of the fields is protected */
/* We make an assumption about the bitmap_t implementation being
static therefore we don't grab locks on read accesses. Anyways,
currently the bitmap_t is used with no write intent only for ASSERTs. */
uint num_free_blocks; /* currently free blocks */
/* Bitmap uses 1KB static data for granularity 64KB and static maximum 512MB */
/* Since we expect only two of these, for now it is ok for users
to have static max rather than dynamically allocating with
exact size - however this field is left last in the structure
in case we do want to save some memory
*/
bitmap_element_t blocks[BITMAP_INDEX(MAX_VMM_HEAP_UNIT_SIZE/VMM_BLOCK_SIZE)];
} vm_heap_t;
/* We keep our heap management structs on the heap for selfprot (case 8074).
* Note that we do have static structs for bootstrapping and we later move
* the data here.
*/
typedef struct _heap_management_t {
/* high-level management */
/* we reserve only a single vm_heap_t for guaranteed allocation,
* we fall back to OS when run out of reservation space */
vm_heap_t vmheap;
heap_t heap;
/* thread-shared heaps: */
thread_units_t global_units;
thread_units_t global_nonpersistent_units;
bool global_heap_writable;
thread_units_t global_unprotected_units;
} heap_management_t;
/* For bootstrapping until we can allocate our real heapmgt (case 8074).
* temp_heapmgt.lock is initialized in vmm_heap_unit_init().
*/
static heap_management_t temp_heapmgt;
static heap_management_t *heapmgt = &temp_heapmgt; /* initial value until alloced */
static bool vmm_heap_exited = false; /* FIXME: used only to thwart stack_free from trying,
should change the interface for the last stack
*/
static inline
uint
vmm_addr_to_block(vm_heap_t *vmh, vm_addr_t p)
{
ASSERT(CHECK_TRUNCATE_TYPE_uint((p - vmh->start_addr) / VMM_BLOCK_SIZE));
return (uint) ((p - vmh->start_addr) / VMM_BLOCK_SIZE);
}
static inline
vm_addr_t
vmm_block_to_addr(vm_heap_t *vmh, uint block)
{
ASSERT(block >=0 && block < vmh->num_blocks);
return (vm_addr_t)(vmh->start_addr + block*VMM_BLOCK_SIZE);
}
static bool
vmm_in_same_block(vm_addr_t p1, vm_addr_t p2)
{
return vmm_addr_to_block(&heapmgt->vmheap, p1) ==
vmm_addr_to_block(&heapmgt->vmheap, p2);
}
#if defined(DEBUG) && defined(INTERNAL)
static void
vmm_dump_map(vm_heap_t *vmh)
{
uint i;
bitmap_element_t *b = vmh->blocks;
uint bitmap_size = vmh->num_blocks;
uint last_i = 0;
bool is_used = bitmap_test(b, 0) == 0;
LOG(GLOBAL, LOG_HEAP, 3, "vmm_dump_map("PFX")\n", vmh);
/* raw dump first - if you really want binary dump use windbg's dyd */
DOLOG(3, LOG_HEAP, {
dump_buffer_as_bytes(GLOBAL, b,
BITMAP_INDEX(bitmap_size)*sizeof(bitmap_element_t),
DUMP_RAW|DUMP_ADDRESS);
});
LOG(GLOBAL, LOG_HEAP, 1, "\nvmm_dump_map("PFX") virtual regions\n", vmh);
#define VMM_DUMP_MAP_LOG(i, last_i) \
LOG(GLOBAL, LOG_HEAP, 1, PFX"-"PFX" size=%d %s\n", vmm_block_to_addr(vmh, last_i), \
vmm_block_to_addr(vmh, i-1) + VMM_BLOCK_SIZE - 1, \
(i-last_i)*VMM_BLOCK_SIZE, \
is_used ? "reserved" : "free");
for (i=0; i < bitmap_size; i++) {
/* start counting at free/used boundaries */
if (is_used != (bitmap_test(b, i) == 0)) {
VMM_DUMP_MAP_LOG(i, last_i);
is_used = (bitmap_test(b, i) == 0);
last_i = i;
}
}
VMM_DUMP_MAP_LOG(bitmap_size, last_i);
}
#endif /* DEBUG */
void
print_vmm_heap_data(file_t outf)
{
mutex_lock(&heapmgt->vmheap.lock);
print_file(outf, "VM heap: addr range "PFX"--"PFX", # free blocks %d\n",
heapmgt->vmheap.start_addr, heapmgt->vmheap.end_addr,
heapmgt->vmheap.num_free_blocks);
mutex_unlock(&heapmgt->vmheap.lock);
}
static inline
void
vmm_heap_initialize_unusable(vm_heap_t *vmh)
{
vmh->start_addr = vmh->end_addr = NULL;
vmh->num_free_blocks = vmh->num_blocks = 0;
}
static
void
vmm_heap_unit_init(vm_heap_t *vmh, size_t size)
{
ptr_uint_t preferred;
heap_error_code_t error_code;
ASSIGN_INIT_LOCK_FREE(vmh->lock, vmh_lock);
size = ALIGN_FORWARD(size, VMM_BLOCK_SIZE);
ASSERT(size <= MAX_VMM_HEAP_UNIT_SIZE);
vmh->alloc_size = size;
if (size == 0) {
vmm_heap_initialize_unusable(&heapmgt->vmheap);
return;
}
/* Out of 32 bits = 12 bits are page offset, windows wastes 4 more
* since its allocation base is 64KB, and if we want to stay
* safely in say 0x20000000-0x2fffffff we're left with only 12
* bits of randomness - which may be too little. On the other
* hand changing any of the lower 16 bits will make our bugs
* non-deterministic. */
/* Make sure we don't waste the lower bits from our random number */
preferred = (DYNAMO_OPTION(vm_base)
+ get_random_offset(DYNAMO_OPTION(vm_max_offset)/VMM_BLOCK_SIZE)
*VMM_BLOCK_SIZE);
preferred = ALIGN_FORWARD(preferred, VMM_BLOCK_SIZE);
/* overflow check: w/ vm_base shouldn't happen so debug-only check */
ASSERT(!POINTER_OVERFLOW_ON_ADD(preferred, size));
/* let's assume a single chunk is sufficient to reserve */
vmh->start_addr = NULL;
#ifdef X64
if ((byte *)preferred < heap_allowable_region_start ||
(byte *)preferred + size > heap_allowable_region_end) {
error_code = HEAP_ERROR_NOT_AT_PREFERRED;
} else {
#endif
vmh->start_addr = os_heap_reserve((void*)preferred, size, &error_code,
true/*+x*/);
LOG(GLOBAL, LOG_HEAP, 1,
"vmm_heap_unit_init preferred="PFX" got start_addr="PFX"\n",
preferred, vmh->start_addr);
#ifdef X64
}
#endif
while (vmh->start_addr == NULL && DYNAMO_OPTION(vm_allow_not_at_base)) {
/* Since we prioritize low-4GB or near-app over -vm_base, we do not
* syslog or assert here
*/
/* need extra size to ensure alignment */
vmh->alloc_size = size + VMM_BLOCK_SIZE;
#ifdef X64
/* PR 215395, make sure allocation satisfies heap reachability contraints */
vmh->alloc_start = os_heap_reserve_in_region
((void *)ALIGN_FORWARD(heap_allowable_region_start, PAGE_SIZE),
(void *)ALIGN_BACKWARD(heap_allowable_region_end, PAGE_SIZE),
size + VMM_BLOCK_SIZE, &error_code,
true/*+x*/);
#else
vmh->alloc_start = (heap_pc)
os_heap_reserve(NULL, size + VMM_BLOCK_SIZE, &error_code, true/*+x*/);
#endif
vmh->start_addr = (heap_pc) ALIGN_FORWARD(vmh->alloc_start, VMM_BLOCK_SIZE);
LOG(GLOBAL, LOG_HEAP, 1, "vmm_heap_unit_init unable to allocate at preferred="
PFX" letting OS place sz=%dM addr="PFX" \n",
preferred, size/(1024*1024), vmh->start_addr);
if (vmh->alloc_start == NULL && DYNAMO_OPTION(vm_allow_smaller)) {
/* Just a little smaller might fit */
size_t sub = (size_t) ALIGN_FORWARD(size/16, 1024*1024);
SYSLOG_INTERNAL_WARNING_ONCE("Full size vmm heap allocation failed");
if (size > sub)
size -= sub;
else
break;
} else
break;
}
#ifdef X64
/* ensure future out-of-block heap allocations are reachable from this allocation */
if (vmh->start_addr != NULL) {
ASSERT(vmh->start_addr >= heap_allowable_region_start &&
!POINTER_OVERFLOW_ON_ADD(vmh->start_addr, size) &&
vmh->start_addr + size <= heap_allowable_region_end);
request_region_be_heap_reachable(vmh->start_addr, size);
}
#endif
if (vmh->start_addr == 0) {
vmm_heap_initialize_unusable(vmh);
/* we couldn't even reserve initial virtual memory - we're out of luck */
/* XXX case 7373: make sure we tag as a potential
* interoperability issue, in staging mode we should probably
* get out from the process since we haven't really started yet
*/
report_low_on_memory(OOM_INIT, error_code);
ASSERT_NOT_REACHED();
}
vmh->end_addr = vmh->start_addr + size;
ASSERT_TRUNCATE(vmh->num_blocks, uint, size / VMM_BLOCK_SIZE);
vmh->num_blocks = (uint) (size / VMM_BLOCK_SIZE);
vmh->num_free_blocks = vmh->num_blocks;
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_unit_init ["PFX","PFX") total=%d free=%d\n",
vmh->start_addr, vmh->end_addr, vmh->num_blocks, vmh->num_free_blocks);
/* make sure static bitmap_t size is properly aligned on block boundaries */
ASSERT(ALIGNED(MAX_VMM_HEAP_UNIT_SIZE, VMM_BLOCK_SIZE));
bitmap_initialize_free(vmh->blocks, vmh->num_blocks);
DOLOG(1, LOG_HEAP, {
vmm_dump_map(vmh);
});
ASSERT(bitmap_check_consistency(vmh->blocks, vmh->num_blocks, vmh->num_free_blocks));
}
static
void
vmm_heap_unit_exit(vm_heap_t *vmh)
{
LOG(GLOBAL, LOG_HEAP, 1, "vmm_heap_unit_exit ["PFX","PFX") total=%d free=%d\n",
vmh->start_addr, vmh->end_addr, vmh->num_blocks, vmh->num_free_blocks);
/* we assume single thread in DR at this point */
DELETE_LOCK(vmh->lock);
if (vmh->start_addr == NULL)
return;
DOLOG(1, LOG_HEAP, { vmm_dump_map(vmh); });
ASSERT(bitmap_check_consistency(vmh->blocks,
vmh->num_blocks, vmh->num_free_blocks));
ASSERT(vmh->num_blocks * VMM_BLOCK_SIZE ==
(ptr_uint_t)(vmh->end_addr - vmh->start_addr));
/* In case there are no tombstones we can just free the unit and
* that is what we'll do, otherwise it will stay up forever.
*/
if (vmh->num_free_blocks == vmh->num_blocks) {
heap_error_code_t error_code;
os_heap_free(vmh->alloc_start, vmh->alloc_size, &error_code);
ASSERT(error_code == HEAP_ERROR_SUCCESS);
} else {
/* FIXME: doing nothing for now - we only care about this in
* detach scenarios where we should try to clean up from the
* virtual address space
*/
}
vmm_heap_initialize_unusable(vmh);
}
/* Returns whether within the region we reserved from the OS for doling
* out internally via our vm_heap_t; asserts that the address was also
* logically reserved within the vm_heap_t.
*/
static
bool
vmm_is_reserved_unit(vm_heap_t *vmh, vm_addr_t p, size_t size)
{
size = ALIGN_FORWARD(size, VMM_BLOCK_SIZE);
if (p < vmh->start_addr || vmh->end_addr < p/*overflow*/ ||
vmh->end_addr < (p + size))
return false;
ASSERT(CHECK_TRUNCATE_TYPE_uint(size/VMM_BLOCK_SIZE));
ASSERT(bitmap_are_reserved_blocks(vmh->blocks, vmh->num_blocks,
vmm_addr_to_block(vmh, p),
(uint)size/VMM_BLOCK_SIZE));
return true;
}
/* Returns whether entirely within the region we reserved from the OS for doling
* out internally via our vm_heap_t
*/
bool
is_vmm_reserved_address(byte *pc, size_t size)
{
ASSERT(heapmgt != NULL);
/* Case 10293: we don't call vmm_is_reserved_unit to avoid its
* assert, which we want to maintain for callers only dealing with
* DR-allocated addresses, while this routine is called w/ random
* addresses
*/
return (heapmgt != NULL && heapmgt->vmheap.start_addr != NULL &&
pc >= heapmgt->vmheap.start_addr &&
!POINTER_OVERFLOW_ON_ADD(pc, size) &&
(pc + size) <= heapmgt->vmheap.end_addr);
}
void
get_vmm_heap_bounds(byte **heap_start/*OUT*/, byte **heap_end/*OUT*/)
{
ASSERT(heapmgt != NULL);
ASSERT(heap_start != NULL && heap_end != NULL);
*heap_start = heapmgt->vmheap.start_addr;
*heap_end = heapmgt->vmheap.end_addr;
}
/* i#774: eventually we'll split vmheap from vmcode. For now, vmcode queries
* refer to the single vmheap reservation.
*/
byte *
vmcode_get_start(void)
{
byte *start, *end;
get_vmm_heap_bounds(&start, &end);
return start;
}
byte *
vmcode_get_end(void)
{
byte *start, *end;
get_vmm_heap_bounds(&start, &end);
return end;
}
byte *
vmcode_unreachable_pc(void)
{
ptr_uint_t start, end;
get_vmm_heap_bounds((byte **)&start, (byte **)&end);
if (start > INT_MAX)
return NULL;
else
return (byte *) PTR_UINT_MINUS_1;
}
bool
rel32_reachable_from_vmcode(byte *tgt)
{
byte *vmcode_start = vmcode_get_start();
byte *vmcode_end = vmcode_get_end();
ptr_int_t new_offs = (tgt > vmcode_start) ? (tgt - vmcode_start) : (vmcode_end - tgt);
/* Beyond-vmm-reservation allocs are handled b/c those are subject to the
* reachability constraints we set up on every new reservation, including
* the initial vm_reserve.
*/
return REL32_REACHABLE_OFFS(new_offs);
}
/* Reservations here are done with VMM_BLOCK_SIZE alignment
* (e.g. 64KB) but the caller is not forced to request at that
* alignment. We explicitly synchronize reservations and decommits
* within the vm_heap_t.
* Returns NULL if the VMMHeap is full or too fragmented to satisfy
* the request.
*/
static vm_addr_t
vmm_heap_reserve_blocks(vm_heap_t *vmh, size_t size_in)
{
vm_addr_t p;
uint request;
uint first_block;
size_t size;
size = ALIGN_FORWARD(size_in, VMM_BLOCK_SIZE);
ASSERT_TRUNCATE(request, uint, size/VMM_BLOCK_SIZE);
request = (uint) size/VMM_BLOCK_SIZE;
LOG(GLOBAL, LOG_HEAP, 2,
"vmm_heap_reserve_blocks: size=%d => %d in blocks=%d free_blocks~=%d\n",
size_in, size, request, vmh->num_free_blocks);
mutex_lock(&vmh->lock);
if (vmh->num_free_blocks < request) {
mutex_unlock(&vmh->lock);
return NULL;
}
first_block = bitmap_allocate_blocks(vmh->blocks, vmh->num_blocks, request);
if (first_block != BITMAP_NOT_FOUND) {
vmh->num_free_blocks -= request;
}
mutex_unlock(&vmh->lock);
if (first_block != BITMAP_NOT_FOUND) {
p = vmm_block_to_addr(vmh, first_block);
STATS_ADD_PEAK(vmm_vsize_used, size);
STATS_ADD_PEAK(vmm_vsize_blocks_used, request);
STATS_ADD_PEAK(vmm_vsize_wasted, size - size_in);
DOSTATS({
if (request > 1) {
STATS_INC(vmm_multi_block_allocs);
STATS_ADD(vmm_multi_blocks, request);
}
});
} else {
p = NULL;
}
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_reserve_blocks: size=%d blocks=%d p="PFX"\n",
size, request, p);
DOLOG(5, LOG_HEAP, { vmm_dump_map(vmh); });
return p;
}
/* We explicitly synchronize reservations and decommits within the vm_heap_t.
* Update bookkeeping information about the freed region.
*/
static void
vmm_heap_free_blocks(vm_heap_t *vmh, vm_addr_t p, size_t size_in)
{
uint first_block = vmm_addr_to_block(vmh, p);
uint request;
size_t size;
size = ALIGN_FORWARD(size_in, VMM_BLOCK_SIZE);
ASSERT_TRUNCATE(request, uint, size/VMM_BLOCK_SIZE);
request = (uint) size/VMM_BLOCK_SIZE;
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_free_blocks: size=%d blocks=%d p="PFX"\n",
size, request, p);
mutex_lock(&vmh->lock);
bitmap_free_blocks(vmh->blocks, vmh->num_blocks, first_block, request);
vmh->num_free_blocks += request;
mutex_unlock(&vmh->lock);
ASSERT(vmh->num_free_blocks <= vmh->num_blocks);
STATS_SUB(vmm_vsize_used, size);
STATS_SUB(vmm_vsize_blocks_used, request);
STATS_SUB(vmm_vsize_wasted, size - size_in);
}
/* This is the proper interface for the rest of heap.c to the os_heap_* functions */
/* place all the local-scope static vars (from DO_THRESHOLD) into .fspdata to avoid
* protection changes */
START_DATA_SECTION(FREQ_PROTECTED_SECTION, "w");
static bool
at_reset_at_vmm_limit()
{
return
(DYNAMO_OPTION(reset_at_vmm_percent_free_limit) != 0 &&
100 * heapmgt->vmheap.num_free_blocks <
DYNAMO_OPTION(reset_at_vmm_percent_free_limit) * heapmgt->vmheap.num_blocks) ||
(DYNAMO_OPTION(reset_at_vmm_free_limit) != 0 &&
heapmgt->vmheap.num_free_blocks * VMM_BLOCK_SIZE <
DYNAMO_OPTION(reset_at_vmm_free_limit));
}
/* Reserve virtual address space without committing swap space for it */
static vm_addr_t
vmm_heap_reserve(size_t size, heap_error_code_t *error_code, bool executable)
{
vm_addr_t p;
/* should only be used on sizable aligned pieces */
ASSERT(size > 0 && ALIGNED(size, PAGE_SIZE));
ASSERT(!OWN_MUTEX(&reset_pending_lock));
if (DYNAMO_OPTION(vm_reserve)) {
/* FIXME: should we make this an external option? */
if (INTERNAL_OPTION(vm_use_last) ||
(DYNAMO_OPTION(switch_to_os_at_vmm_reset_limit) && at_reset_at_vmm_limit())) {
DO_ONCE({
if (DYNAMO_OPTION(reset_at_switch_to_os_at_vmm_limit))
schedule_reset(RESET_ALL);
DOCHECK(1, {
if (!INTERNAL_OPTION(vm_use_last)) {
ASSERT_CURIOSITY(false && "running low on vm reserve");
}
});
/* FIXME - for our testing would be nice to have some release build
* notification of this ... */
});
DODEBUG(ever_beyond_vmm = true;);
#ifdef X64
/* PR 215395, make sure allocation satisfies heap reachability contraints */
p = os_heap_reserve_in_region
((void *)ALIGN_FORWARD(heap_allowable_region_start, PAGE_SIZE),
(void *)ALIGN_BACKWARD(heap_allowable_region_end, PAGE_SIZE),
size, error_code, executable);
/* ensure future heap allocations are reachable from this allocation */
if (p != NULL)
request_region_be_heap_reachable(p, size);
#else
p = os_heap_reserve(NULL, size, error_code, executable);
#endif
if (p != NULL)
return p;
LOG(GLOBAL, LOG_HEAP, 1, "vmm_heap_reserve: failed "PFX"\n",
*error_code);
}
if (at_reset_at_vmm_limit()) {
/* We're running low on our reservation, trigger a reset */
if (schedule_reset(RESET_ALL)) {
STATS_INC(reset_low_vmm_count);
DO_THRESHOLD_SAFE(DYNAMO_OPTION(report_reset_vmm_threshold),
FREQ_PROTECTED_SECTION,
{/* < max - nothing */},
{/* >= max */
/* FIXME - do we want to report more then once to give some idea of
* how much thrashing there is? */
DO_ONCE({
SYSLOG_CUSTOM_NOTIFY(SYSLOG_WARNING, MSG_LOW_ON_VMM_MEMORY, 2,
"Potentially thrashing on low virtual "
"memory resetting.", get_application_name(),
get_application_pid());
/* want QA to notice */
ASSERT_CURIOSITY(false && "vmm heap limit reset thrashing");
});
});
}
}
p = vmm_heap_reserve_blocks(&heapmgt->vmheap, size);
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_reserve: size=%d p="PFX"\n",
size, p);
if (p)
return p;
DO_ONCE({
DODEBUG({ out_of_vmheap_once = true; });
if (!INTERNAL_OPTION(skip_out_of_vm_reserve_curiosity)) {
/* this maybe unsafe for early services w.r.t. case 666 */
SYSLOG_INTERNAL_WARNING("Out of vmheap reservation - reserving %dKB."
"Falling back onto OS allocation", size/1024);
ASSERT_CURIOSITY(false && "Out of vmheap reservation");
}
/* This actually-out trigger is only trying to help issues like a
* thread-private configuration being a memory hog (and thus we use up
* our reserve). Reset needs memory, and this is asynchronous, so no
* guarantees here anyway (the app may have already reserved all memory
* beyond our reservation, see sqlsrvr.exe and cisvc.exe for ex.) which is
* why we have -reset_at_vmm_threshold to make reset success more likely. */
if (DYNAMO_OPTION(reset_at_vmm_full)) {
schedule_reset(RESET_ALL);
}
});
}
/* if we fail to allocate from our reservation we fall back to the OS */
DODEBUG(ever_beyond_vmm = true;);
#ifdef X64
/* PR 215395, make sure allocation satisfies heap reachability contraints */
p = os_heap_reserve_in_region
((void *)ALIGN_FORWARD(heap_allowable_region_start, PAGE_SIZE),
(void *)ALIGN_BACKWARD(heap_allowable_region_end, PAGE_SIZE),
size, error_code, executable);
/* ensure future heap allocations are reachable from this allocation */
if (p != NULL)
request_region_be_heap_reachable(p, size);
#else
p = os_heap_reserve(NULL, size, error_code, executable);
#endif
return p;
}
/* Commit previously reserved pages, returns false when out of memory
* This is here just to complement the vmm interface, in fact it is
* almost an alias for os_heap_commit. (If we had strict types then
* here we'd convert a vm_addr_t into a heap_pc.)
*/
static inline bool
vmm_heap_commit(vm_addr_t p, size_t size, uint prot, heap_error_code_t *error_code)
{
bool res = os_heap_commit(p, size, prot, error_code);
size_t commit_used, commit_limit;
ASSERT(!OWN_MUTEX(&reset_pending_lock));
if ((DYNAMO_OPTION(reset_at_commit_percent_free_limit) != 0 ||
DYNAMO_OPTION(reset_at_commit_free_limit) != 0) &&
os_heap_get_commit_limit(&commit_used, &commit_limit)) {
size_t commit_left = commit_limit - commit_used;
ASSERT(commit_used <= commit_limit);
/* FIXME - worry about overflow in the multiplies below? With 4kb pages isn't
* an issue till 160GB of committable memory. */
if ((DYNAMO_OPTION(reset_at_commit_free_limit) != 0 &&
commit_left < DYNAMO_OPTION(reset_at_commit_free_limit) / PAGE_SIZE) ||
(DYNAMO_OPTION(reset_at_commit_percent_free_limit) != 0 &&
100 * commit_left <
DYNAMO_OPTION(reset_at_commit_percent_free_limit) * commit_limit)) {
/* Machine is getting low on memory, trigger a reset */
/* FIXME - if we aren't the ones hogging committed memory (rougue app) then
* do we want a version of reset that doesn't de-commit our already grabbed
* memory to avoid someone else stealing it (or perhaps keep just a minimal
* level to ensure we make some progress)? */
/* FIXME - the commit limit is for the whole system; we have no good way of
* telling if we're running in a job and if so what the commit limit for the
* job is. */
/* FIXME - if a new process is started under dr while the machine is already
* past the threshold we will just spin resetting here and not make any
* progress, may be better to only reset when we have a reasonable amount of
* non-persistent memory to free (so that we can at least make some progress
* before resetting again). */
/* FIXME - the threshold is calculated at the current page file size, but
* it's possible that the pagefile is expandable (dependent on disk space of
* course) and thus we're preventing a potentially beneficial (to us)
* upsizing of the pagefile here. See "HKLM\SYSTEM\CCS\ControlSession /
* Manager\Memory Management" for the initial/max size of the various page
* files (query SystemPafefileInformation only gets you the current size). */
/* xref case 345 on fixmes (and link to wiki discussion) */
if (schedule_reset(RESET_ALL)) {
STATS_INC(reset_low_commit_count);
DO_THRESHOLD_SAFE(DYNAMO_OPTION(report_reset_commit_threshold),
FREQ_PROTECTED_SECTION,
{/* < max - nothing */},
{/* >= max */
/* FIXME - do we want to report more then once to give some idea of
* how much thrashing there is? */
DO_ONCE({
SYSLOG_CUSTOM_NOTIFY(SYSLOG_WARNING,
MSG_LOW_ON_COMMITTABLE_MEMORY, 2,
"Potentially thrashing on low committable "
"memory resetting.", get_application_name(),
get_application_pid());
/* want QA to notice */
ASSERT_CURIOSITY(false && "commit limit reset thrashing");
});
});
}
}
}
if (!res &&
DYNAMO_OPTION(oom_timeout) != 0) {
DEBUG_DECLARE(heap_error_code_t old_error_code = *error_code;)
ASSERT(old_error_code != HEAP_ERROR_SUCCESS);
/* check whether worth retrying */
if (!os_heap_systemwide_overcommit(*error_code)) {
/* FIXME: we should check whether current process is the hog */
/* unless we have used the memory, there is still a
* miniscule chance another thread will free up some or
* will attempt suicide, so could retry even if current
* process has a leak */
ASSERT_NOT_IMPLEMENTED(false);
/* retry */
}
SYSLOG_INTERNAL_WARNING("vmm_heap_commit oom: timeout and retry");
/* let's hope a memory hog dies in the mean time */
os_timeout(DYNAMO_OPTION(oom_timeout));
res = os_heap_commit(p, size, prot, error_code);
DODEBUG({
if (res) {
SYSLOG_INTERNAL_WARNING("vmm_heap_commit retried, got away! old="PFX" new="PFX"\n",
old_error_code, *error_code);
} else {
SYSLOG_INTERNAL_WARNING("vmm_heap_commit retrying, no luck. old="PFX" new="PFX"\n",
old_error_code, *error_code);
}
});
}
return res;
}
/* back to normal section */
END_DATA_SECTION()
/* Free previously reserved and possibly committed memory. Check if
* it is within the memory managed by the virtual memory manager we
* only decommit back to the OS, and we remove the vmm reservation.
* Keep in mind that this can be called on units that are not fully
* committed, e.g. guard pages are added to this - as long as the
* os_heap_decommit interface can handle this we're OK
*/
static void
vmm_heap_free(vm_addr_t p, size_t size, heap_error_code_t *error_code)
{
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_free: size=%d p="PFX" is_reserved=%d\n",
size, p, vmm_is_reserved_unit(&heapmgt->vmheap, p, size));
/* the memory doesn't have to be within our VM reserve if it
was allocated as an extra OS call when if we ran out */
if (DYNAMO_OPTION(vm_reserve)) {
if (vmm_is_reserved_unit(&heapmgt->vmheap, p, size)) {
os_heap_decommit(p, size, error_code);
vmm_heap_free_blocks(&heapmgt->vmheap, p, size);
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_free: freed size=%d p="PFX"\n",
size, p);
return;
} else {
/* FIXME: check if this is stack_free getting in the way, then ignore it */
/* FIXME: could do this by overriding the meaning of the vmheap fields
after cleanup to a different combination that start_pc = end_pc = NULL
*/
/* FIXME: see vmm_heap_unit_exit for the current stack_free problem */
if (vmm_heap_exited) {
*error_code = HEAP_ERROR_SUCCESS;
return;
}
}
}
os_heap_free(p, size, error_code);
}
static void
vmm_heap_decommit(vm_addr_t p, size_t size, heap_error_code_t *error_code)
{
LOG(GLOBAL, LOG_HEAP, 2, "vmm_heap_decommit: size=%d p="PFX" is_reserved=%d\n",
size, p, vmm_is_reserved_unit(&heapmgt->vmheap, p, size));
os_heap_decommit(p, size, error_code);
/* nothing to be done to vmm blocks */
}
/* Caller is required to handle thread synchronization and to update dynamo vm areas.
* size must be PAGE_SIZE-aligned.
* Returns NULL if fails to allocate memory!
*/
static void *
vmm_heap_alloc(size_t size, uint prot, heap_error_code_t *error_code)
{
vm_addr_t p = vmm_heap_reserve(size, error_code, TEST(MEMPROT_EXEC, prot));
if (!p)
return NULL; /* out of reserved memory */
if (!vmm_heap_commit(p, size, prot, error_code))
return NULL; /* out of committed memory */
return p;
}
/* set reachability constraints before loading any client libs */
void
vmm_heap_init_constraints()
{
#ifdef X64
/* add reachable regions before we allocate the heap, xref PR 215395 */
/* i#774, i#901: we no longer need the DR library nor ntdll.dll to be
* reachable by the vmheap reservation. But, for -heap_in_lower_4GB,
* we must call request_region_be_heap_reachable() up front.
*/
if (DYNAMO_OPTION(heap_in_lower_4GB))
request_region_be_heap_reachable((byte *)(ptr_uint_t)0x80000000/*middle*/, 1);
else if (DYNAMO_OPTION(vm_base_near_app)) {
/* Required for STATIC_LIBRARY: must be near app b/c clients are there.
* Non-static: still a good idea for fewer rip-rel manglings.
* Asking for app base means we'll prefer before the app, which
* has less of an impact on its heap.
*/
app_pc base = get_application_base();
request_region_be_heap_reachable(base, get_application_end() - base);
} else {
/* It seems silly to let the 1st client lib set the region, so we give
* -vm_base priority.
*/
request_region_be_heap_reachable
((byte *)DYNAMO_OPTION(vm_base), DYNAMO_OPTION(vm_size));
}
/* XXX: really we should iterate and try other options: right now we'll
* just fail if we run out of space. E.g., if the app is quite large, we might
* fit the client near, and then our vm reservation could fail and at that
* point we'd just abort. We need to restructure the code to allow
* iterating over the client lib loads and vm reservation at once.
*/
#endif /* X64 */
}
/* virtual memory manager initialization */
void
vmm_heap_init()
{
IF_WINDOWS(ASSERT(VMM_BLOCK_SIZE == OS_ALLOC_GRANULARITY));
if (DYNAMO_OPTION(vm_reserve)) {
vmm_heap_unit_init(&heapmgt->vmheap, DYNAMO_OPTION(vm_size));
}
}
void
vmm_heap_exit()
{
/* virtual memory manager exit */
if (DYNAMO_OPTION(vm_reserve)) {
/* FIXME: we have three regions that are not explicitly
* deallocated current stack, init stack, global_do_syscall
*/
DOCHECK(1, {
uint perstack =
ALIGN_FORWARD_UINT(dynamo_options.stack_size +
(dynamo_options.guard_pages ? (2*PAGE_SIZE) : 0),
VMM_BLOCK_SIZE) /
VMM_BLOCK_SIZE;
uint unfreed_blocks = perstack * 1 /* initstack */ +
/* current stack */
perstack * ((IF_WINDOWS_ELSE(doing_detach, false)
IF_APP_EXPORTS(|| dr_api_exit)) ? 0 : 1);
/* FIXME: on detach arch_thread_exit should explicitly mark as
left behind all TPCs needed so then we can assert even for
detach
*/
ASSERT(IF_WINDOWS(doing_detach || ) /* not deterministic when detaching */
heapmgt->vmheap.num_free_blocks == heapmgt->vmheap.num_blocks
- unfreed_blocks ||
/* >=, not ==, b/c if we hit the vmm limit the cur dstack
* could be outside of vmm (i#1164)
*/
(ever_beyond_vmm && heapmgt->vmheap.num_free_blocks >=
heapmgt->vmheap.num_blocks - unfreed_blocks));
});
/* FIXME: On process exit we are currently executing off a
* stack in this region so we cannot free the whole allocation.
* FIXME: Any tombstone allocations will have to use a
* different interface than the generic heap_mmap() which is
* sometimes used to leave things behind. FIXME: Currently
* we'll leave behind the whole vm unit if any tombstones are
* left - which in fact is always the case, no matter whether
* thread private code needs to be left or not.
* global_do_syscall 32 byte allocation should be part of our
* dll and won't have to be left.
* The current stack is the main problem because it is later
* cleaned up in cleanup_and_terminate by calling stack_free which
* in turn gets all the way to vmm_heap_free. Therefore we add an
* explicit test for vmm_heap_exited, so that we can otherwise free
* bookkeeping information and delete the lock now.
* Potential solution to most of these problems is to have
* cleanup_and_terminate call vmm_heap_exit when cleaning up
* the process, or to just leave the vm mapping behind and
* simply pass a different argument to stack_free.
*/
vmm_heap_unit_exit(&heapmgt->vmheap);
vmm_heap_exited = true;
}
}
/* checks for compatibility among heap options, returns true if
* modified the value of any options to make them compatible
*/
bool
heap_check_option_compatibility()
{
bool ret = false;
ret = check_param_bounds(&dynamo_options.vm_size, MIN_VMM_HEAP_UNIT_SIZE,
MAX_VMM_HEAP_UNIT_SIZE, "vm_size")
|| ret;
#ifdef INTERNAL
/* if max_heap_unit_size is too small you may get a funny message
* "initial_heap_unit_size must be >= 8229 and <= 4096" but in
* release build we will take the min and then complain about
* max_heap_unit_size and set it to the min also, so it all works
* out w/o needing an extra check() call.
*/
/* case 7626: don't short-circuit checks, as later ones may be needed */
ret = check_param_bounds(&dynamo_options.initial_heap_unit_size,
/* if have units smaller than a page we end up
* allocating 64KB chunks for "oversized" units
* for just about every alloc! so round up to
* at least a page.
*/
ALIGN_FORWARD(UNITOVERHEAD + 1, PAGE_SIZE),
HEAP_UNIT_MAX_SIZE, "initial_heap_unit_size")
|| ret;
ret = check_param_bounds(&dynamo_options.initial_global_heap_unit_size,
ALIGN_FORWARD(UNITOVERHEAD + 1, PAGE_SIZE),
HEAP_UNIT_MAX_SIZE, "initial_global_heap_unit_size")
|| ret;
ret = check_param_bounds(&dynamo_options.max_heap_unit_size,
MAX(HEAP_UNIT_MIN_SIZE, GLOBAL_UNIT_MIN_SIZE),
INT_MAX, "max_heap_unit_size")
|| ret;
#endif
return ret;
}
/* thread-shared initialization that should be repeated after a reset */
void
heap_reset_init()
{
if (DYNAMO_OPTION(enable_reset)) {
threadunits_init(GLOBAL_DCONTEXT, &heapmgt->global_nonpersistent_units,
GLOBAL_UNIT_MIN_SIZE);
}
}
/* initialization */
void
heap_init()
{
int i;
uint prev_sz = 0;
LOG(GLOBAL, LOG_TOP|LOG_HEAP, 2, "Heap bucket sizes are:\n");
/* make sure we'll preserve alignment */
ASSERT(ALIGNED(HEADER_SIZE, HEAP_ALIGNMENT));
/* make sure free list pointers will fit */
ASSERT(BLOCK_SIZES[0] >= sizeof(heap_pc*));
/* since sizes depend on size of structs, make sure they're in order */
for (i = 0; i < BLOCK_TYPES; i++) {
ASSERT(BLOCK_SIZES[i] > prev_sz);
/* we assume all of our heap allocs are aligned */
ASSERT(i == BLOCK_TYPES-1 || ALIGNED(BLOCK_SIZES[i], HEAP_ALIGNMENT));
prev_sz = BLOCK_SIZES[i];
LOG(GLOBAL, LOG_TOP|LOG_HEAP, 2, "\t%d bytes\n", BLOCK_SIZES[i]);
}
/* we assume writes to some static vars are atomic,
* i.e., the vars don't cross cache lines. they shouldn't since
* they should all be 4-byte-aligned in the data segment.
* FIXME: ensure that release build aligns ok?
* I would be quite surprised if static vars were not 4-byte-aligned!
*/
ASSERT(ALIGN_BACKWARD(&heap_exiting, CACHE_LINE_SIZE()) ==
ALIGN_BACKWARD(&heap_exiting + 1, CACHE_LINE_SIZE()));
ASSERT(ALIGN_BACKWARD(&heap_unit_lock.owner, CACHE_LINE_SIZE()) ==
ALIGN_BACKWARD(&heap_unit_lock.owner + 1, CACHE_LINE_SIZE()));
/* For simplicity we go through our normal heap mechanism to allocate
* our post-init heapmgt struct
*/
ASSERT(heapmgt == &temp_heapmgt);
heapmgt->global_heap_writable = true;
threadunits_init(GLOBAL_DCONTEXT, &heapmgt->global_units, GLOBAL_UNIT_MIN_SIZE);
heapmgt = HEAP_TYPE_ALLOC(GLOBAL_DCONTEXT, heap_management_t, ACCT_MEM_MGT, PROTECTED);
memset(heapmgt, 0, sizeof(*heapmgt));
ASSERT(sizeof(temp_heapmgt) == sizeof(*heapmgt));
memcpy(heapmgt, &temp_heapmgt, sizeof(temp_heapmgt));
threadunits_init(GLOBAL_DCONTEXT, &heapmgt->global_unprotected_units,
GLOBAL_UNIT_MIN_SIZE);
heap_reset_init();
#ifdef WINDOWS
/* PR 250294: As part of 64-bit hook work, hook reachability was addressed
* using landing pads (see win32/callback.c for more explanation). Landing
* pad areas are a type of special heap, so they should be initialized
* during heap init.
* Each landing pad area has its own allocation pointer, so they shouldn't
* be merged automatically.
*/
VMVECTOR_ALLOC_VECTOR(landing_pad_areas, GLOBAL_DCONTEXT,
VECTOR_SHARED | VECTOR_NEVER_MERGE,
landing_pad_areas_lock);
#endif
}
/* need to not remove from vmareas on process exit -- vmareas has already exited! */
static void
really_free_unit(heap_unit_t *u)
{
STATS_SUB(heap_capacity, UNIT_COMMIT_SIZE(u));
STATS_ADD(heap_reserved_only, (stats_int_t)(UNIT_COMMIT_SIZE(u) - UNIT_RESERVED_SIZE(u)));
/* remember that u itself is inside unit, not separately allocated */
release_guarded_real_memory((vm_addr_t)u, UNIT_RESERVED_SIZE(u),
false/*do not update DR areas now*/, true);
}
/* Free all thread-shared state not critical to forward progress;
* heap_reset_init() will be called before continuing.
*/
void
heap_reset_free()
{
heap_unit_t *u, *next_u;
/* FIXME: share some code w/ heap_exit -- currently only called by reset */
ASSERT(DYNAMO_OPTION(enable_reset));
/* we must grab this lock before heap_unit_lock to avoid rank
* order violations when freeing
*/
dynamo_vm_areas_lock();
/* for combining stats into global_units we need this lock
* FIXME: remove if we go to separate stats sum location
*/
DODEBUG({ acquire_recursive_lock(&global_alloc_lock); });
acquire_recursive_lock(&heap_unit_lock);
LOG(GLOBAL, LOG_HEAP, 1, "Pre-reset, global heap unit stats:\n");
/* FIXME: free directly rather than putting on dead list first */
threadunits_exit(&heapmgt->global_nonpersistent_units, GLOBAL_DCONTEXT);
/* free all dead units */
u = heapmgt->heap.dead;
while (u != NULL) {
next_u = u->next_global;
LOG(GLOBAL, LOG_HEAP, 1, "\tfreeing dead unit "PFX"-"PFX" [-"PFX"]\n",
u, UNIT_COMMIT_END(u), UNIT_RESERVED_END(u));
RSTATS_DEC(heap_num_free);
really_free_unit(u);
u = next_u;
}
heapmgt->heap.dead = NULL;
heapmgt->heap.num_dead = 0;
release_recursive_lock(&heap_unit_lock);
DODEBUG({ release_recursive_lock(&global_alloc_lock); });
dynamo_vm_areas_unlock();
}
/* atexit cleanup */
void
heap_exit()
{
heap_unit_t *u, *next_u;
heap_management_t *temp;
heap_exiting = true;
/* FIXME: we shouldn't need either lock if executed last */
dynamo_vm_areas_lock();
acquire_recursive_lock(&heap_unit_lock);
#ifdef WINDOWS
release_landing_pad_mem(); /* PR 250294 */
#endif
LOG(GLOBAL, LOG_HEAP, 1, "Global unprotected heap unit stats:\n");
threadunits_exit(&heapmgt->global_unprotected_units, GLOBAL_DCONTEXT);
if (DYNAMO_OPTION(enable_reset)) {
LOG(GLOBAL, LOG_HEAP, 1, "Global nonpersistent heap unit stats:\n");
threadunits_exit(&heapmgt->global_nonpersistent_units, GLOBAL_DCONTEXT);
}
/* Now we need to go back to the static struct to clean up */
ASSERT(heapmgt != &temp_heapmgt);
memcpy(&temp_heapmgt, heapmgt, sizeof(temp_heapmgt));
temp = heapmgt;
heapmgt = &temp_heapmgt;
HEAP_TYPE_FREE(GLOBAL_DCONTEXT, temp, heap_management_t, ACCT_MEM_MGT, PROTECTED);
LOG(GLOBAL, LOG_HEAP, 1, "Global heap unit stats:\n");
threadunits_exit(&heapmgt->global_units, GLOBAL_DCONTEXT);
/* free heap for all unfreed units */
LOG(GLOBAL, LOG_HEAP, 1, "Unfreed units:\n");
u = heapmgt->heap.units;
while (u != NULL) {
next_u = u->next_global;
LOG(GLOBAL, LOG_HEAP, 1, "\tfreeing live unit "PFX"-"PFX" [-"PFX"]\n",
u, UNIT_COMMIT_END(u), UNIT_RESERVED_END(u));
RSTATS_DEC(heap_num_live);
really_free_unit(u);
u = next_u;
}
heapmgt->heap.units = NULL;
u = heapmgt->heap.dead;
while (u != NULL) {
next_u = u->next_global;
LOG(GLOBAL, LOG_HEAP, 1, "\tfreeing dead unit "PFX"-"PFX" [-"PFX"]\n",
u, UNIT_COMMIT_END(u), UNIT_RESERVED_END(u));
RSTATS_DEC(heap_num_free);
really_free_unit(u);
u = next_u;
}
heapmgt->heap.dead = NULL;
release_recursive_lock(&heap_unit_lock);
dynamo_vm_areas_unlock();
DELETE_RECURSIVE_LOCK(heap_unit_lock);
DELETE_RECURSIVE_LOCK(global_alloc_lock);
#ifdef X64
DELETE_LOCK(request_region_be_heap_reachable_lock);
#endif
}
/* FIXME:
* detect if the app is who we're fighting for memory, if so, don't
* free memory, else the app will just keep grabbing more.
* need a test for hitting 2GB (or 3GB!) user mode limit.
*/
static void
heap_low_on_memory()
{
/* free some memory! */
heap_unit_t *u, *next_u;
size_t freed = 0;
LOG(GLOBAL, LOG_CACHE|LOG_STATS, 1,
"heap_low_on_memory: about to free dead list units\n");
/* WARNING: this routine is called at arbitrary allocation failure points,
* so we have to be careful what locks we grab
* However, no allocation site can hold a lock weaker in rank than
* heap_unit_lock, b/c it could deadlock on the allocation itself!
* So we're safe.
*/
/* must grab this lock prior to heap_unit_lock if affecting DR vm areas
* this is recursive so ok if we ran out of memory while holding DR vm area lock
*/
ASSERT(safe_to_allocate_or_free_heap_units());
dynamo_vm_areas_lock();
acquire_recursive_lock(&heap_unit_lock);
u = heapmgt->heap.dead;
while (u != NULL) {
next_u = u->next_global;
freed += UNIT_COMMIT_SIZE(u);
/* FIXME: if out of committed pages only, could keep our reservations */
LOG(GLOBAL, LOG_HEAP, 1, "\tfreeing dead unit "PFX"-"PFX" [-"PFX"]\n",
u, UNIT_COMMIT_END(u), UNIT_RESERVED_END(u));
RSTATS_DEC(heap_num_free);
really_free_unit(u);
u = next_u;
heapmgt->heap.num_dead--;
}
heapmgt->heap.dead = NULL;
release_recursive_lock(&heap_unit_lock);
dynamo_vm_areas_unlock();
LOG(GLOBAL, LOG_CACHE|LOG_STATS, 1,
"heap_low_on_memory: freed %d KB\n", freed/1024);
/* FIXME: we don't keep a list of guard pages, which we may decide to throw
* out or compact at this time.
*/
/* FIXME: should also fix up the allocator to look in other free lists
* of sizes larger than asked for, we may have plenty of memory available
* in other lists! see comments in common_heap_alloc
*/
}
static const char*
get_oom_source_name(oom_source_t source)
{
/* currently only single character codenames,
* (still as a string though)
*/
const char *code_name = "?";
switch (source) {
case OOM_INIT : code_name = "I"; break;
case OOM_RESERVE : code_name = "R"; break;
case OOM_COMMIT : code_name = "C"; break;
case OOM_EXTEND : code_name = "E"; break;
default:
ASSERT_NOT_REACHED();
}
return code_name;
}
static bool
silent_oom_for_process(oom_source_t source)
{
if (TESTANY(OOM_COMMIT|OOM_EXTEND, source) &&
!IS_STRING_OPTION_EMPTY(silent_commit_oom_list)) {
bool onlist;
const char *process_name = get_short_name(get_application_name());
string_option_read_lock();
onlist = check_filter_with_wildcards(DYNAMO_OPTION(silent_commit_oom_list),
process_name);
string_option_read_unlock();
if (onlist) {
SYSLOG_INTERNAL_WARNING("not reporting last words of executable %s",
process_name);
return true;
}
}
return false;
}
/* oom_source_t identifies the action we were taking, os_error_code is
* the returned value from the last system call - opaque at this OS
* independent layer.
*/
static void
report_low_on_memory(oom_source_t source, heap_error_code_t os_error_code)
{
if (TESTANY(DYNAMO_OPTION(silent_oom_mask), source)
|| silent_oom_for_process(source)) {
SYSLOG_INTERNAL_WARNING("Mostly silent OOM: %s "PFX".\n",
get_oom_source_name(source), os_error_code);
/* still produce an ldmp for internal use */
if (TEST(DUMPCORE_OUT_OF_MEM_SILENT, DYNAMO_OPTION(dumpcore_mask)))
os_dump_core("Out of memory, silently aborting program.");
} else {
const char *oom_source_code = get_oom_source_name(source);
char status_hex[19]; /* FIXME: for 64bit hex need 16+NULL */
/* note 0x prefix added by the syslog */
snprintf(status_hex,
BUFFER_SIZE_ELEMENTS(status_hex), PFX, /* FIXME: 32bit */
os_error_code);
NULL_TERMINATE_BUFFER(status_hex);
/* SYSLOG first */
SYSLOG_CUSTOM_NOTIFY(SYSLOG_CRITICAL, MSG_OUT_OF_MEMORY, 4,
"Out of memory. Program aborted.",
get_application_name(), get_application_pid(),
oom_source_code, status_hex
);
/* FIXME: case 7306 can't specify arguments in SYSLOG_CUSTOM_NOTIFY */
SYSLOG_INTERNAL_WARNING("OOM Status: %s %s", oom_source_code, status_hex);
/* FIXME: case 7296 - ldmp even if we have decided not to produce an event above */
if (TEST(DUMPCORE_OUT_OF_MEM, DYNAMO_OPTION(dumpcore_mask)))
os_dump_core("Out of memory, aborting program.");
/* passing only status code to XML where we should have a stack dump and callstack */
report_diagnostics("Out of memory", status_hex, NO_VIOLATION_BAD_INTERNAL_STATE);
}
os_terminate(NULL, TERMINATE_PROCESS);
ASSERT_NOT_REACHED();
}
/* update statistics for committed memory, and add to vm_areas */
static inline void
account_for_memory(void *p, size_t size, uint prot, bool add_vm, bool image
_IF_DEBUG(const char *comment))
{
STATS_ADD_PEAK(memory_capacity, size);
/* case 3045: areas inside the vmheap reservation are not added to the list
* for clients that use DR-allocated memory, we have get_memory_info()
* query from the OS to see inside
*/
if (vmm_is_reserved_unit(&heapmgt->vmheap, p, size)) {
return;
}
if (add_vm) {
add_dynamo_vm_area(p, ((app_pc)p) + size, prot, image _IF_DEBUG(comment));
} else {
/* due to circular dependencies bet vmareas and global heap we do not call
* add_dynamo_vm_area here, instead we indicate that something has changed
*/
mark_dynamo_vm_areas_stale();
/* NOTE: 'prot' info is lost about this region, but is needed in
* heap_vmareas_synch_units to update all_memory_areas. Currently
* heap_create_unit is the only place that passes 'false' with prot rw-.
*/
ASSERT(TESTALL(MEMPROT_READ | MEMPROT_WRITE, prot));
}
}
/* remove_vm MUST be false iff this is heap memory, which is updated separately */
static void
update_dynamo_areas_on_release(app_pc start, app_pc end, bool remove_vm)
{
if (!vm_areas_exited && !heap_exiting) { /* avoid problems when exiting */
/* case 3045: areas inside the vmheap reservation are not added to the list
* for clients that use DR-allocated memory, we have get_memory_info()
* query from the OS to see inside
*/
if (vmm_is_reserved_unit(&heapmgt->vmheap, start, end - start)) {
return;
}
if (remove_vm) {
remove_dynamo_vm_area(start, end);
} else {
/* due to cyclic dependencies bet heap and vmareas we cannot remove incrementally.
* the pending set is protected by the same lock needed to synch the vm areas,
* so we will never mis-identify free memory as DR memory.
*/
mark_dynamo_vm_areas_stale();
dynamo_areas_pending_remove = true;
}
}
}
bool
lockwise_safe_to_allocate_memory()
{
/* check whether it's safe to hold a lock that normally can be held
* for memory allocation -- i.e., check whether we hold the
* global_alloc_lock
*/
return !self_owns_recursive_lock(&global_alloc_lock);
}
/* we indirect all os memory requests through here so we have a central place
* to handle the out-of-memory condition.
* add_vm MUST be false iff this is heap memory, which is updated separately.
*/
static void *
get_real_memory(size_t size, uint prot, bool add_vm _IF_DEBUG(const char *comment))
{
void *p;
heap_error_code_t error_code;
/* must round up to page sizes, else vmm_heap_alloc assert triggers */
size = ALIGN_FORWARD(size, PAGE_SIZE);
/* memory alloc/dealloc and updating DR list must be atomic */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
p = vmm_heap_alloc(size, prot, &error_code);
if (p == NULL) {
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory -- cannot reserve or "
"commit %dKB. Trying to recover.", size/1024);
/* we should be ok here, shouldn't come in here holding global_alloc_lock
* or heap_unit_lock w/o first having grabbed DR areas lock
*/
ASSERT(safe_to_allocate_or_free_heap_units());
heap_low_on_memory();
fcache_low_on_memory();
/* try again
* FIXME: have more sophisticated strategy of freeing a little, then getting
* more drastic with each subsequent failure
* FIXME: can only free live fcache units for current thread w/ current
* impl...should we wait a while and try again if out of memory, hoping
* other threads have freed some?!?!
*/
p = vmm_heap_alloc(size, prot, &error_code);
if (p == NULL) {
report_low_on_memory(OOM_RESERVE, error_code);
}
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory -- but still alive after "
"emergency free.");
}
account_for_memory(p, size, prot, add_vm, false _IF_DEBUG(comment));
dynamo_vm_areas_unlock();
return p;
}
static void
release_memory_and_update_areas(app_pc p, size_t size, bool decommit, bool remove_vm)
{
heap_error_code_t error_code;
/* these two operations need to be atomic wrt DR area updates */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
/* ref case 3035, we must remove from dynamo_areas before we free in case
* we end up allocating memory in the process of removing the area
* (we don't want to end up getting the memory we just freed since that
* would lead to errors in the list when we finally did remove it)
*/
update_dynamo_areas_on_release(p, p + size, remove_vm);
if (decommit)
vmm_heap_decommit(p, size, &error_code);
else
vmm_heap_free(p, size, &error_code);
ASSERT(error_code == HEAP_ERROR_SUCCESS);
dynamo_vm_areas_unlock();
}
/* remove_vm MUST be false iff this is heap memory, which is updated separately */
static void
release_real_memory(void *p, size_t size, bool remove_vm)
{
/* must round up to page sizes for vmm_heap_free */
size = ALIGN_FORWARD(size, PAGE_SIZE);
release_memory_and_update_areas((app_pc)p, size, false/*free*/, remove_vm);
/* avoid problem w/ being called by cleanup_and_terminate after dynamo_process_exit */
DOSTATS({
if (!dynamo_exited_log_and_stats)
STATS_SUB(memory_capacity, size);
});
}
static void
extend_commitment(vm_addr_t p, size_t size, uint prot,
bool initial_commit)
{
heap_error_code_t error_code;
ASSERT(ALIGNED(p, PAGE_SIZE));
size = ALIGN_FORWARD(size, PAGE_SIZE);
if (!vmm_heap_commit(p, size, prot, &error_code)) {
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory - cannot extend commit "
"%dKB. Trying to recover.", size/1024);
heap_low_on_memory();
fcache_low_on_memory();
/* see low-memory ideas in get_real_memory */
if (!vmm_heap_commit(p, size, prot, &error_code)) {
report_low_on_memory(initial_commit ? OOM_COMMIT : OOM_EXTEND,
error_code);
}
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory in extend - still alive "
"after emergency free.");
}
}
/* a wrapper around get_real_memory that adds a guard page on each side of the requested unit.
* These should consume only uncommitted virtual address and should not use any physical memory.
* add_vm MUST be false iff this is heap memory, which is updated separately.
*/
static vm_addr_t
get_guarded_real_memory(size_t reserve_size, size_t commit_size, uint prot,
bool add_vm, bool guarded _IF_DEBUG(const char *comment))
{
vm_addr_t p;
uint guard_size = PAGE_SIZE;
heap_error_code_t error_code;
ASSERT(reserve_size >= commit_size);
if (!guarded || !dynamo_options.guard_pages) {
if (reserve_size == commit_size)
return get_real_memory(reserve_size, prot, add_vm _IF_DEBUG(comment));
guard_size = 0;
}
reserve_size = ALIGN_FORWARD(reserve_size, PAGE_SIZE);
commit_size = ALIGN_FORWARD(commit_size, PAGE_SIZE);
reserve_size += 2* guard_size; /* add top and bottom guards */
/* memory alloc/dealloc and updating DR list must be atomic */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
p = vmm_heap_reserve(reserve_size, &error_code, TEST(MEMPROT_EXEC, prot));
if (p == NULL) {
/* This is very unlikely to happen - we have to reach at least 2GB reserved memory. */
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory - cannot reserve %dKB. "
"Trying to recover.", reserve_size/1024);
heap_low_on_memory();
fcache_low_on_memory();
p = vmm_heap_reserve(reserve_size, &error_code, TEST(MEMPROT_EXEC, prot));
if (p == NULL) {
report_low_on_memory(OOM_RESERVE, error_code);
}
SYSLOG_INTERNAL_WARNING_ONCE("Out of memory on reserve - but still "
"alive after emergency free.");
}
/* includes guard pages if add_vm -- else, heap_vmareas_synch_units() will
* add guard pages in by assuming one page on each side of every heap unit
* if dynamo_options.guard_pages
*/
account_for_memory((void *)p, reserve_size, prot, add_vm, false _IF_DEBUG(comment));
dynamo_vm_areas_unlock();
STATS_ADD_PEAK(reserved_memory_capacity, reserve_size);
STATS_ADD_PEAK(guard_pages, 2);
p += guard_size;
extend_commitment(p, commit_size, prot, true /* initial commit */);
return p;
}
/* a wrapper around get_release_memory that also frees the guard pages on each side of the requested unit.
* remove_vm MUST be false iff this is heap memory, which is updated separately
*/
static void
release_guarded_real_memory(vm_addr_t p, size_t size, bool remove_vm, bool guarded)
{
if (!guarded || !dynamo_options.guard_pages) {
release_real_memory(p, size, remove_vm);
return;
}
size = ALIGN_FORWARD(size, PAGE_SIZE);
size += PAGE_SIZE * 2; /* add top and bottom guards */
p -= PAGE_SIZE;
release_memory_and_update_areas((app_pc)p, size, false/*free*/, remove_vm);
/* avoid problem w/ being called by cleanup_and_terminate after dynamo_process_exit */
DOSTATS({
if (!dynamo_exited_log_and_stats) {
STATS_SUB(memory_capacity, size);
STATS_SUB(reserved_memory_capacity, size);
STATS_ADD(guard_pages, -2);
}
});
}
/* use heap_mmap to allocate large chunks of executable memory
* it's mainly used to allocate our fcache units
*/
void *
heap_mmap_ex(size_t reserve_size, size_t commit_size, uint prot, bool guarded)
{
/* XXX i#774: when we split vmheap and vmcode, if MEMPROT_EXEC is requested
* here (or this is a call from a client, for reachability
* compatibility), put it in vmcode; else in vmheap.
*/
void *p = get_guarded_real_memory(reserve_size, commit_size, prot, true, guarded
_IF_DEBUG("heap_mmap"));
#ifdef DEBUG_MEMORY
if (TEST(MEMPROT_WRITE, prot))
memset(p, HEAP_ALLOCATED_BYTE, commit_size);
#endif
/* We rely on this for freeing _post_stack in absence of dcontext */
ASSERT(!DYNAMO_OPTION(vm_reserve) ||
!DYNAMO_OPTION(stack_shares_gencode) ||
(ptr_uint_t)p - (guarded ? (GUARD_PAGE_ADJUSTMENT/2) : 0) ==
ALIGN_BACKWARD(p, VMM_BLOCK_SIZE) ||
at_reset_at_vmm_limit());
LOG(GLOBAL, LOG_HEAP, 2, "heap_mmap: %d bytes [/ %d] @ "PFX"\n",
commit_size, reserve_size, p);
STATS_ADD_PEAK(mmap_capacity, commit_size);
STATS_ADD_PEAK(mmap_reserved_only, (reserve_size - commit_size));
return p;
}
/* use heap_mmap to allocate large chunks of executable memory
* it's mainly used to allocate our fcache units
*/
void *
heap_mmap_reserve(size_t reserve_size, size_t commit_size)
{
/* heap_mmap always marks as executable */
return heap_mmap_ex(reserve_size, commit_size,
MEMPROT_EXEC|MEMPROT_READ|MEMPROT_WRITE, true);
}
/* It is up to the caller to ensure commit_size is a page size multiple,
* and that it does not extend beyond the initial reservation.
*/
void
heap_mmap_extend_commitment(void *p, size_t commit_size)
{
extend_commitment(p, commit_size, MEMPROT_EXEC|MEMPROT_READ|MEMPROT_WRITE,
false /*not initial commit*/);
STATS_SUB(mmap_reserved_only, commit_size);
STATS_ADD_PEAK(mmap_capacity, commit_size);
#ifdef DEBUG_MEMORY
memset(p, HEAP_ALLOCATED_BYTE, commit_size);
#endif
}
/* De-commits from a committed region. */
void
heap_mmap_retract_commitment(void *retract_start, size_t decommit_size)
{
heap_error_code_t error_code;
ASSERT(ALIGNED(decommit_size, PAGE_SIZE));
vmm_heap_decommit(retract_start, decommit_size, &error_code);
STATS_ADD(mmap_reserved_only, decommit_size);
STATS_ADD_PEAK(mmap_capacity, -(stats_int_t)decommit_size);
}
/* Allocates executable memory in the same allocation region as this thread's
* stack, to save address space (case 9474).
*/
void *
heap_mmap_reserve_post_stack(dcontext_t *dcontext,
size_t reserve_size, size_t commit_size)
{
void *p;
byte *stack_reserve_end = NULL;
heap_error_code_t error_code;
size_t available = 0;
uint prot;
bool known_stack = false;
ASSERT(reserve_size > 0 && commit_size < reserve_size);
/* 1.5 * guard page adjustment since we'll share the middle one */
if (DYNAMO_OPTION(stack_size) + reserve_size +
GUARD_PAGE_ADJUSTMENT + GUARD_PAGE_ADJUSTMENT / 2 > VMM_BLOCK_SIZE) {
/* there's not enough room to share the allocation block, stack is too big */
LOG(GLOBAL, LOG_HEAP, 1, "Not enough room to allocate 0x%08x bytes post stack "
"of size 0x%08x\n", reserve_size, DYNAMO_OPTION(stack_size));
return heap_mmap_reserve(reserve_size, commit_size);
}
if (DYNAMO_OPTION(stack_shares_gencode) &&
/* FIXME: we could support this w/o vm_reserve, or when beyond
* the reservation, but we don't bother */
DYNAMO_OPTION(vm_reserve) &&
dcontext != GLOBAL_DCONTEXT && dcontext != NULL) {
stack_reserve_end = dcontext->dstack + GUARD_PAGE_ADJUSTMENT/2;
#if defined(UNIX) && !defined(HAVE_MEMINFO)
prot = 0; /* avoid compiler warning: should only need inside if */
if (!dynamo_initialized) {
/* memory info is not yet set up. since so early we only support
* post-stack if inside vmm (won't be true only for pathological
* tiny vmm sizes)
*/
if (vmm_is_reserved_unit(&heapmgt->vmheap, stack_reserve_end, reserve_size)) {
known_stack = true;
available = reserve_size;
} else
known_stack = false;
} else
#elif defined(UNIX)
/* the all_memory_areas list doesn't keep details inside vmheap */
known_stack = get_memory_info_from_os(stack_reserve_end, NULL,
&available, &prot);
#else
known_stack = get_memory_info(stack_reserve_end, NULL, &available, &prot);
#endif
/* If ever out of vmheap, then may have free space beyond stack,
* which we could support but don't (see FIXME above) */
ASSERT(out_of_vmheap_once ||
(known_stack && available >= reserve_size && prot == 0));
}
if (!known_stack ||
/* if -no_vm_reserve will short-circuit so no vmh deref danger */
!vmm_in_same_block(dcontext->dstack,
/* we do want a guard page at the end */
stack_reserve_end + reserve_size) ||
available < reserve_size) {
ASSERT(!DYNAMO_OPTION(stack_shares_gencode) ||
!DYNAMO_OPTION(vm_reserve) || out_of_vmheap_once);
DOLOG(1, LOG_HEAP, {
if (known_stack && available < reserve_size) {
LOG(GLOBAL, LOG_HEAP, 1,
"heap_mmap_reserve_post_stack: avail %d < needed %d\n",
available, reserve_size);
}
});
STATS_INC(mmap_no_share_stack_region);
return heap_mmap_reserve(reserve_size, commit_size);
}
ASSERT(DYNAMO_OPTION(vm_reserve));
ASSERT(stack_reserve_end != NULL);
prot = MEMPROT_EXEC|MEMPROT_READ|MEMPROT_WRITE;
/* memory alloc/dealloc and updating DR list must be atomic */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
/* We share the stack's end guard page as our start guard page */
if (vmm_is_reserved_unit(&heapmgt->vmheap, stack_reserve_end, reserve_size)) {
/* Memory is already reserved with OS */
p = stack_reserve_end;
} else {
p = os_heap_reserve(stack_reserve_end, reserve_size, &error_code, true/*+x*/);
#ifdef X64
/* ensure future heap allocations are reachable from this allocation
* (this will also verify that this region meets reachability requirements) */
if (p != NULL)
request_region_be_heap_reachable(p, reserve_size);
#endif
if (p == NULL) {
ASSERT_NOT_REACHED();
LOG(GLOBAL, LOG_HEAP, 1,
"heap_mmap_reserve_post_stack: reserve failed "PFX"\n", error_code);
dynamo_vm_areas_unlock();
STATS_INC(mmap_no_share_stack_region);
return heap_mmap_reserve(reserve_size, commit_size);
}
ASSERT(error_code == HEAP_ERROR_SUCCESS);
}
if (!vmm_heap_commit(p, commit_size, prot, &error_code)) {
ASSERT_NOT_REACHED();
LOG(GLOBAL, LOG_HEAP, 1, "heap_mmap_reserve_post_stack: commit failed "PFX"\n",
error_code);
if (!vmm_is_reserved_unit(&heapmgt->vmheap, stack_reserve_end, reserve_size)) {
os_heap_free(p, reserve_size, &error_code);
ASSERT(error_code == HEAP_ERROR_SUCCESS);
}
dynamo_vm_areas_unlock();
STATS_INC(mmap_no_share_stack_region);
return heap_mmap_reserve(reserve_size, commit_size);
}
account_for_memory(p, reserve_size, prot, true/*add now*/, false
_IF_DEBUG("heap_mmap_reserve_post_stack"));
dynamo_vm_areas_unlock();
/* We rely on this for freeing in absence of dcontext */
ASSERT((ptr_uint_t)p - GUARD_PAGE_ADJUSTMENT/2 !=
ALIGN_BACKWARD(p, VMM_BLOCK_SIZE));
#ifdef DEBUG_MEMORY
memset(p, HEAP_ALLOCATED_BYTE, commit_size);
#endif
LOG(GLOBAL, LOG_HEAP, 2, "heap_mmap w/ stack: %d bytes [/ %d] @ "PFX"\n",
commit_size, reserve_size, p);
STATS_ADD_PEAK(mmap_capacity, commit_size);
STATS_ADD_PEAK(mmap_reserved_only, (reserve_size - commit_size));
STATS_INC(mmap_share_stack_region);
return p;
}
/* De-commits memory that was allocated in the same allocation region as this
* thread's stack (case 9474).
*/
void
heap_munmap_post_stack(dcontext_t *dcontext, void *p, size_t reserve_size)
{
/* We would require a valid dcontext and compare to the stack reserve end,
* but on detach we have no dcontext, so we instead use block alignment.
*/
DOCHECK(1, {
if (dcontext != NULL && dcontext != GLOBAL_DCONTEXT &&
DYNAMO_OPTION(vm_reserve) && DYNAMO_OPTION(stack_shares_gencode)) {
bool at_stack_end = (p == dcontext->dstack + GUARD_PAGE_ADJUSTMENT/2);
bool at_block_start = ((ptr_uint_t)p - GUARD_PAGE_ADJUSTMENT/2 ==
ALIGN_BACKWARD(p, VMM_BLOCK_SIZE));
ASSERT((at_stack_end && !at_block_start) ||
(!at_stack_end && at_block_start));
}
});
if (!DYNAMO_OPTION(vm_reserve) ||
!DYNAMO_OPTION(stack_shares_gencode) ||
(ptr_uint_t)p - GUARD_PAGE_ADJUSTMENT/2 == ALIGN_BACKWARD(p, VMM_BLOCK_SIZE)) {
heap_munmap(p, reserve_size);
} else {
/* Detach makes it a pain to pass in the commit size so
* we use the reserve size, which works fine.
*/
release_memory_and_update_areas((app_pc)p, reserve_size, true/*decommit*/,
true/*update now*/);
LOG(GLOBAL, LOG_HEAP, 2, "heap_munmap_post_stack: %d bytes @ "PFX"\n",
reserve_size, p);
STATS_SUB(mmap_capacity, reserve_size);
STATS_SUB(mmap_reserved_only, reserve_size);
}
}
/* use heap_mmap to allocate large chunks of executable memory
* it's mainly used to allocate our fcache units
*/
void *
heap_mmap(size_t size)
{
return heap_mmap_reserve(size, size);
}
/* free memory-mapped storage */
void
heap_munmap_ex(void *p, size_t size, bool guarded)
{
#ifdef DEBUG_MEMORY
/* can't set to HEAP_UNALLOCATED_BYTE since really not in our address
* space anymore */
#endif
release_guarded_real_memory((vm_addr_t)p, size, true/*update DR areas immediately*/,
guarded);
DOSTATS({
/* avoid problem w/ being called by cleanup_and_terminate after dynamo_process_exit */
if (!dynamo_exited_log_and_stats) {
LOG(GLOBAL, LOG_HEAP, 2, "heap_munmap: %d bytes @ "PFX"\n", size, p);
STATS_SUB(mmap_capacity, size);
STATS_SUB(mmap_reserved_only, size);
}
});
}
/* free memory-mapped storage */
void
heap_munmap(void *p, size_t size)
{
heap_munmap_ex(p, size, true/*guarded*/);
}
#ifdef STACK_GUARD_PAGE
# define STACK_GUARD_PAGES 1
#endif
/* use stack_alloc to build a stack -- it returns TOS
* For STACK_GUARD_PAGE, it also marks the bottom STACK_GUARD_PAGES==1
* to detect overflows when used.
*/
void *
stack_alloc(size_t size)
{
void *p;
/* we reserve and commit at once for now
* FIXME case 2330: commit-on-demand could allow larger max sizes w/o
* hurting us in the common case
*/
p = get_guarded_real_memory(size, size, MEMPROT_READ|MEMPROT_WRITE, true, true
_IF_DEBUG("stack_alloc"));
#ifdef DEBUG_MEMORY
memset(p, HEAP_ALLOCATED_BYTE, size);
#endif
#ifdef STACK_GUARD_PAGE
/* mark the bottom page non-accessible to trap stack overflow */
/* NOTE: the guard page should be included in the total memory requested */
# ifdef WINDOWS
mark_page_as_guard((byte *)p + ((STACK_GUARD_PAGES - 1) * PAGE_SIZE));
# else
/* FIXME: make no access, not just no write -- and update signal.c to
* look at reads and not just writes -- though unwritable is nearly as good
*/
# if defined(CLIENT_INTERFACE) || defined(STANDALONE_UNIT_TEST)
if (!standalone_library)
# endif
make_unwritable(p, STACK_GUARD_PAGES * PAGE_SIZE);
# endif
#endif
STATS_ADD(stack_capacity, size);
STATS_MAX(peak_stack_capacity, stack_capacity);
/* stack grows from high to low */
return (void *) ((ptr_uint_t)p + size);
}
/* free stack storage */
void
stack_free(void *p, size_t size)
{
if (size == 0)
size = DYNAMORIO_STACK_SIZE;
p = (void *) ((vm_addr_t)p - size);
release_guarded_real_memory((vm_addr_t)p, size, true/*update DR areas immediately*/,
true);
DOSTATS({
if (!dynamo_exited_log_and_stats)
STATS_SUB(stack_capacity, size);
});
}
#ifdef STACK_GUARD_PAGE
/* only checks initstack and current dcontext
* does not check any dstacks on the callback stack (win32) */
bool
is_stack_overflow(dcontext_t *dcontext, byte *sp)
{
/* ASSUMPTION: size of stack is DYNAMORIO_STACK_SIZE = dynamo_options.stack_size
* currently sideline violates that for a thread stack
* but all dstacks and initstack should be this size
*/
byte *bottom = dcontext->dstack - DYNAMORIO_STACK_SIZE;
/* see if in bottom guard page of dstack */
if (sp >= bottom && sp < bottom + (STACK_GUARD_PAGES * PAGE_SIZE))
return true;
/* now check the initstack */
bottom = initstack - DYNAMORIO_STACK_SIZE;
if (sp >= bottom && sp < bottom + (STACK_GUARD_PAGES * PAGE_SIZE))
return true;
return false;
}
#endif
byte *
map_file(file_t f, size_t *size INOUT, uint64 offs, app_pc addr, uint prot,
map_flags_t map_flags)
{
byte *view;
/* memory alloc/dealloc and updating DR list must be atomic */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
view = os_map_file(f, size, offs, addr, prot, map_flags);
if (view != NULL) {
STATS_ADD_PEAK(file_map_capacity, *size);
account_for_memory((void *)view, *size, prot, true/*add now*/, true/*image*/
_IF_DEBUG("map_file"));
}
dynamo_vm_areas_unlock();
return view;
}
bool
unmap_file(byte *map, size_t size)
{
bool success;
ASSERT(map != NULL && ALIGNED(map, PAGE_SIZE));
size = ALIGN_FORWARD(size, PAGE_SIZE);
/* memory alloc/dealloc and updating DR list must be atomic */
dynamo_vm_areas_lock(); /* if already hold lock this is a nop */
success = os_unmap_file(map, size);
if (success) {
/* Only update the all_memory_areas on success.
* It should still be atomic to the outside observers.
*/
update_dynamo_areas_on_release(map, map+size, true/*remove now*/);
STATS_SUB(file_map_capacity, size);
}
dynamo_vm_areas_unlock();
return success;
}
/* We cannot incrementally keep dynamo vm area list up to date due to
* circular dependencies bet vmareas and global heap (trust me, I've tried
* to support it with reentrant routines and recursive locks, the hard part
* is getting add_vm_area to be reentrant or to queue up adding areas,
* I think this solution is much more elegant, plus it avoids race conditions
* between DR memory allocation and the vmareas list by ensuring the list
* is up to date at the exact time of each query).
* Instead we on-demand walk the units.
* Freed units can usually be removed incrementally, except when we
* hold the heap_unit_lock when we run out of memory -- when we set
* a flag telling the caller of this routine to remove all heap areas
* from the vm list prior to calling us to add the real ones back in.
* Re-adding everyone is the simplest policy, so we don't have to keep
* track of who's been added.
* The caller is assumed to hold the dynamo vm areas write lock.
*/
void
heap_vmareas_synch_units()
{
heap_unit_t *u, *next;
/* make sure to add guard page on each side, as well */
uint offs = (dynamo_options.guard_pages) ? PAGE_SIZE : 0;
/* we again have circular dependence w/ vmareas if it happens to need a
* new unit in the course of adding these areas, so we use a recursive lock!
* furthermore, we need to own the global lock now, to avoid deadlock with
* another thread who does global_alloc and then needs a new unit!
* which means that the global_alloc lock must be recursive since vmareas
* may need to global_alloc...
*/
/* if chance could own both locks, must grab both now
* always grab global_alloc first, then we won't have deadlocks
*/
acquire_recursive_lock(&global_alloc_lock);
acquire_recursive_lock(&heap_unit_lock);
if (dynamo_areas_pending_remove) {
dynamo_areas_pending_remove = false;
remove_dynamo_heap_areas();
/* When heap units are removed from the dynamo_area, they should be
* marked so. See case 4196.
*/
for (u = heapmgt->heap.units; u != NULL; u = u->next_global)
u->in_vmarea_list = false;
for (u = heapmgt->heap.dead; u != NULL; u = u->next_global)
u->in_vmarea_list = false;
}
for (u = heapmgt->heap.units; u != NULL; u = next) {
app_pc start = (app_pc)u - offs;
/* support un-aligned heap reservation end: PR 415269 (though as
* part of that PR we shouldn't have un-aligned anymore)
*/
app_pc end_align = (app_pc) ALIGN_FORWARD(UNIT_RESERVED_END(u), PAGE_SIZE);
app_pc end = end_align + offs;
/* u can be moved to dead list, so cache the next link; case 4196. */
next = u->next_global;
/* case 3045: areas inside the vmheap reservation are not added to the list */
if (!u->in_vmarea_list && !vmm_is_reserved_unit(&heapmgt->vmheap,
start, end - start)) {
/* case 4196 if next is used by dynamo_vmareas then next
* may become dead if vector is resized, then u should be
* alive and u->next_global should be reset AFTER add */
bool next_may_die =
is_dynamo_area_buffer(UNIT_GET_START_PC(next)); /* keep breaking abstractions */
/* dynamo_areas.buf vector may get resized and u can either
* go to the dead unit list, or it can be released back to
* the OS. We'll mark it as being in vmarea list to avoid
* re-adding when going through dead one's, and we'll mark
* _before_ the potential free. If dynamo_areas.buf is
* freed back to the OS we'll have another iteration in
* update_dynamo_vm_areas() until we get fully
* synchronized, so we don't need to worry about the
* inconsistency.
*/
u->in_vmarea_list = true;
add_dynamo_heap_vm_area(start, end, true, false _IF_DEBUG("heap unit"));
/* NOTE: Since we could mark_dynamo_vm_areas_stale instead of adding to
* it, we may lose prot info about this unit.
* FIXME: Currently, this is done only at one place, which allocates unit
* as MEMPROT_READ | MEMPROT_WRITE. If other places are added, then this
* needs to change.
*/
update_all_memory_areas((app_pc)u, end_align,
MEMPROT_READ | MEMPROT_WRITE,
DR_MEMTYPE_DATA); /* unit */
if (offs != 0) {
/* guard pages */
update_all_memory_areas((app_pc)u - offs, (app_pc)u, MEMPROT_NONE,
DR_MEMTYPE_DATA);
update_all_memory_areas(end_align, end, MEMPROT_NONE,
DR_MEMTYPE_DATA);
}
if (next_may_die) {
STATS_INC(num_vmareas_resize_synch);
/* if next was potentially on dead row, then current
* should still be live and point to the next live
*/
next = u->next_global;
}
}
}
for (u = heapmgt->heap.dead; u != NULL; u = next) {
app_pc start = (app_pc)u - offs;
/* support un-aligned heap reservation end: PR 415269 (though as
* part of that PR we shouldn't have un-aligned anymore)
*/
app_pc end_align = (app_pc) ALIGN_FORWARD(UNIT_RESERVED_END(u), PAGE_SIZE);
app_pc end = end_align + offs;
/* u can be moved to live list, so cache the next link; case 4196. */
next = u->next_global;
/* case 3045: areas inside the vmheap reservation are not added to the list */
if (!u->in_vmarea_list && !vmm_is_reserved_unit(&heapmgt->vmheap,
start, end - start)) {
u->in_vmarea_list = true;
add_dynamo_heap_vm_area(start, end, true, false _IF_DEBUG("dead heap unit"));
update_all_memory_areas((app_pc)u, end_align,
MEMPROT_READ | MEMPROT_WRITE,
DR_MEMTYPE_DATA); /* unit */
if (offs != 0) {
/* guard pages */
update_all_memory_areas(start, (app_pc)u, MEMPROT_NONE,
DR_MEMTYPE_DATA);
update_all_memory_areas(end_align, end, MEMPROT_NONE,
DR_MEMTYPE_DATA);
}
/* case 4196 if next was put back on live list for
* dynamo_areas.buf vector, then next will no longer be a
* valid iterator over dead list
*/
if (is_dynamo_area_buffer(UNIT_GET_START_PC(next))) { /* keep breaking abstractions */
STATS_INC(num_vmareas_resize_synch);
ASSERT_NOT_TESTED();
next = u->next_global;
}
}
}
release_recursive_lock(&heap_unit_lock);
release_recursive_lock(&global_alloc_lock);
}
/* shared between global and global_unprotected */
static void *
common_global_heap_alloc(thread_units_t *tu, size_t size HEAPACCT(which_heap_t which))
{
void *p;
acquire_recursive_lock(&global_alloc_lock);
p = common_heap_alloc(tu, size HEAPACCT(which));
release_recursive_lock(&global_alloc_lock);
if (p == NULL) {
/* circular dependence solution: we need to hold DR lock before
* global alloc lock -- so we back out, grab it, and retry
*/
dynamo_vm_areas_lock();
acquire_recursive_lock(&global_alloc_lock);
p = common_heap_alloc(tu, size HEAPACCT(which));
release_recursive_lock(&global_alloc_lock);
dynamo_vm_areas_unlock();
}
ASSERT(p != NULL);
return p;
}
/* shared between global and global_unprotected */
static void
common_global_heap_free(thread_units_t *tu, void *p, size_t size HEAPACCT(which_heap_t which))
{
bool ok;
if (p == NULL) {
ASSERT(false && "attempt to free NULL");
return;
}
acquire_recursive_lock(&global_alloc_lock);
ok = common_heap_free(tu, p, size HEAPACCT(which));
release_recursive_lock(&global_alloc_lock);
if (!ok) {
/* circular dependence solution: we need to hold DR lock before
* global alloc lock -- so we back out, grab it, and retry
*/
dynamo_vm_areas_lock();
acquire_recursive_lock(&global_alloc_lock);
ok = common_heap_free(tu, p, size HEAPACCT(which));
release_recursive_lock(&global_alloc_lock);
dynamo_vm_areas_unlock();
}
ASSERT(ok);
}
/* these functions use the global heap instead of a thread's heap: */
void *
global_heap_alloc(size_t size HEAPACCT(which_heap_t which))
{
void *p = common_global_heap_alloc(&heapmgt->global_units, size HEAPACCT(which));
ASSERT(p != NULL);
LOG(GLOBAL, LOG_HEAP, 6, "\nglobal alloc: "PFX" (%d bytes)\n", p, size);
return p;
}
void
global_heap_free(void *p, size_t size HEAPACCT(which_heap_t which))
{
common_global_heap_free(&heapmgt->global_units, p, size HEAPACCT(which));
LOG(GLOBAL, LOG_HEAP, 6, "\nglobal free: "PFX" (%d bytes)\n", p, size);
}
/* reallocate area
allocates new_num elements of element_size
if ptr is NULL acts like global_heap_alloc,
copies an old_num elements of given size in the new area */
/* FIXME: do a heap_realloc and a special_heap_realloc too */
void *
global_heap_realloc(void *ptr, size_t old_num, size_t new_num, size_t element_size
HEAPACCT(which_heap_t which))
{
void *new_area = global_heap_alloc(new_num * element_size HEAPACCT(which));
if (ptr) {
memcpy(new_area, ptr, (old_num < new_num ? old_num : new_num) * element_size);
global_heap_free(ptr, old_num * element_size HEAPACCT(which));
}
return new_area;
}
/* size does not include guard pages (if any) and is reserved, but only
* DYNAMO_OPTION(heap_commit_increment) is committed up front
*/
static heap_unit_t *
heap_create_unit(thread_units_t *tu, size_t size, bool must_be_new)
{
heap_unit_t *u = NULL, *dead = NULL, *prev_dead = NULL;
bool new_unit = false;
/* we do not restrict size to unit max as we have to make larger-than-max
* units for oversized requests
*/
/* modifying heap list and DR areas must be atomic, and must grab
* DR area lock before heap_unit_lock
*/
ASSERT(safe_to_allocate_or_free_heap_units());
dynamo_vm_areas_lock();
/* take from dead list if possible */
acquire_recursive_lock(&heap_unit_lock);
/* FIXME: need to unprotect units that we're going to perform
* {next,prev}_global assignments too -- but need to know whether
* to re-protect -- do all at once, or each we need? add a writable
* flag to heap_unit_t?
*/
if (!must_be_new) {
for (dead = heapmgt->heap.dead;
dead != NULL && UNIT_RESERVED_SIZE(dead) < size;
prev_dead = dead, dead = dead->next_global)
;
}
if (dead != NULL) {
if (prev_dead == NULL)
heapmgt->heap.dead = dead->next_global;
else
prev_dead->next_global = dead->next_global;
u = dead;
heapmgt->heap.num_dead--;
RSTATS_DEC(heap_num_free);
release_recursive_lock(&heap_unit_lock);
LOG(GLOBAL, LOG_HEAP, 2,
"Re-using dead heap unit: "PFX"-"PFX" %d KB (need %d KB)\n",
u, ((byte*)u)+size, UNIT_RESERVED_SIZE(u)/1024, size/1024);
} else {
size_t commit_size = DYNAMO_OPTION(heap_commit_increment);
release_recursive_lock(&heap_unit_lock); /* do not hold while asking for memory */
/* create new unit */
ASSERT(commit_size <= size);
u = (heap_unit_t *)
get_guarded_real_memory(size, commit_size, MEMPROT_READ|MEMPROT_WRITE,
false, true _IF_DEBUG(""));
new_unit = true;
/* FIXME: handle low memory conditions by freeing units, + fcache units? */
ASSERT(u);
LOG(GLOBAL, LOG_HEAP, 2, "New heap unit: "PFX"-"PFX"\n", u, ((byte*)u)+size);
/* u is kept at top of unit itself, so displace start pc */
u->start_pc = (heap_pc) (((ptr_uint_t)u) + sizeof(heap_unit_t));
u->end_pc = ((heap_pc)u) + commit_size;
u->reserved_end_pc = ((heap_pc)u) + size;
u->in_vmarea_list = false;
STATS_ADD(heap_capacity, commit_size);
STATS_MAX(peak_heap_capacity, heap_capacity);
/* FIXME: heap sizes are not always page-aligned so stats will be off */
STATS_ADD_PEAK(heap_reserved_only, (u->reserved_end_pc - u->end_pc));
}
RSTATS_ADD_PEAK(heap_num_live, 1);
u->cur_pc = u->start_pc;
u->next_local = NULL;
DODEBUG({
u->id = tu->num_units;
tu->num_units++;
});
acquire_recursive_lock(&heap_unit_lock);
u->next_global = heapmgt->heap.units;
if (heapmgt->heap.units != NULL)
heapmgt->heap.units->prev_global = u;
u->prev_global = NULL;
heapmgt->heap.units = u;
release_recursive_lock(&heap_unit_lock);
dynamo_vm_areas_unlock();
#ifdef DEBUG_MEMORY
DOCHECK(CHKLVL_MEMFILL,
memset(u->start_pc, HEAP_UNALLOCATED_BYTE, u->end_pc - u->start_pc););
#endif
return u;
}
/* dcontext only used to determine whether a global unit or not */
static void
heap_free_unit(heap_unit_t *unit, dcontext_t *dcontext)
{
heap_unit_t *u, *prev_u;
#ifdef DEBUG_MEMORY
/* Unit should already be set to all HEAP_UNALLOCATED by the individual
* frees and the free list cleanup, verify. */
/* NOTE - this assert fires if any memory in the unit wasn't freed. This
* would include memory allocated ACCT_TOMBSTONE (which we don't currently
* use). Using ACCT_TOMBSTONE is dangerous since we will still free the
* unit here (say at proc or thread exit) even if there are ACCT_TOMBSTONE
* allocations in it. */
/* Note, this memset check is done only on the special heap unit header,
* not on the unit itself - FIXME: case 10434. Maybe we should embed the
* special heap unit header in the first special heap unit itself. */
/* The hotp_only leak relaxation below is for case 9588 & 9593. */
DOCHECK(CHKLVL_MEMFILL, {
CLIENT_ASSERT(IF_HOTP(hotp_only_contains_leaked_trampoline
(unit->start_pc, unit->end_pc - unit->start_pc) ||)
/* i#157: private loader => system lib allocs come here =>
* they don't always clean up. we have to relax here, but our
* threadunits_exit checks should find all leaks anyway.
*/
heapmgt->global_units.acct.cur_usage[ACCT_LIBDUP] > 0 ||
is_region_memset_to_char(unit->start_pc,
unit->end_pc - unit->start_pc,
HEAP_UNALLOCATED_BYTE)
/* don't assert when client does premature exit as it's
* hard for Extension libs, etc. to clean up in such situations
*/
IF_CLIENT_INTERFACE(|| client_requested_exit),
"memory leak detected");
});
#endif
/* modifying heap list and DR areas must be atomic, and must grab
* DR area lock before heap_unit_lock
*/
ASSERT(safe_to_allocate_or_free_heap_units());
dynamo_vm_areas_lock();
acquire_recursive_lock(&heap_unit_lock);
/* FIXME: need to unprotect units that we're going to perform
* {next,prev}_global assignments too -- but need to know whether
* to re-protect -- do all at once, or each we need? add a writable
* flag to heap_unit_t?
*/
/* remove from live list */
if (unit->prev_global != NULL) {
unit->prev_global->next_global = unit->next_global;
} else
heapmgt->heap.units = unit->next_global;
if (unit->next_global != NULL) {
unit->next_global->prev_global = unit->prev_global;
}
/* prev_global is not used in the dead list */
unit->prev_global = NULL;
RSTATS_DEC(heap_num_live);
/* heuristic: don't keep around more dead units than max(5, 1/4 num threads)
* FIXME: share the policy with the fcache dead unit policy
* also, don't put special larger-than-max units on free list -- though
* we do now have support for doing so (after PR 415269)
*/
if (UNITALLOC(unit) <= HEAP_UNIT_MAX_SIZE &&
(heapmgt->heap.num_dead < 5 ||
heapmgt->heap.num_dead * 4U <= (uint) get_num_threads())) {
/* Keep dead list sorted small-to-large to avoid grabbing large
* when can take small and then needing to allocate when only
* have small left. Helps out with lots of small threads.
*/
for (u = heapmgt->heap.dead, prev_u = NULL;
u != NULL && UNIT_RESERVED_SIZE(u) < UNIT_RESERVED_SIZE(unit);
prev_u = u, u = u->next_global)
;
if (prev_u == NULL) {
unit->next_global = heapmgt->heap.dead;
heapmgt->heap.dead = unit;
} else {
unit->next_global = u;
prev_u->next_global = unit;
}
heapmgt->heap.num_dead++;
release_recursive_lock(&heap_unit_lock);
RSTATS_ADD_PEAK(heap_num_free, 1);
} else {
/* don't need to hold this while freeing since still hold DR areas lock */
release_recursive_lock(&heap_unit_lock);
LOG(GLOBAL, LOG_HEAP, 1, "\tfreeing excess dead unit "PFX"-"PFX" [-"PFX"]\n",
unit, UNIT_COMMIT_END(unit), UNIT_RESERVED_END(unit));
really_free_unit(unit);
}
/* FIXME: shrink lock-held path if we see contention */
dynamo_vm_areas_unlock();
}
#ifdef DEBUG_MEMORY
static heap_unit_t *
find_heap_unit(thread_units_t *tu, heap_pc p, size_t size)
{
/* FIXME (case 6198): this is a perf hit in debug builds. But, we can't use
* a new vmvector b/c of circular dependences. Proposal: use custom data
* field of vm_area_t in dynamo_areas list for heap entries to store a pointer
* to the heap_unit_t struct, and add a backpointer to the owning thread_units_t
* in heap_unit_t. Then have to make sure it's ok lock-wise to query the
* dynamo_areas in the middle of an alloc or a free. It should be but for
* global alloc and free we will have to grab the dynamo_areas lock up front
* every time instead of the very rare times now when we need a new unit.
*/
heap_unit_t *unit;
ASSERT(!POINTER_OVERFLOW_ON_ADD(p, size)); /* should not overflow */
for (unit = tu->top_unit;
unit != NULL && (p < unit->start_pc || p+size > unit->end_pc);
unit = unit->next_local);
return unit;
}
#endif
static void
threadunits_init(dcontext_t *dcontext, thread_units_t *tu, size_t size)
{
int i;
DODEBUG({
tu->num_units = 0;
});
tu->top_unit = heap_create_unit(tu, size - GUARD_PAGE_ADJUSTMENT,
false/*can reuse*/);
tu->cur_unit = tu->top_unit;
tu->dcontext = dcontext;
tu->writable = true;
#ifdef HEAP_ACCOUNTING
memset(&tu->acct, 0, sizeof(tu->acct));
#endif
for (i=0; i<BLOCK_TYPES; i++)
tu->free_list[i] = NULL;
}
#ifdef HEAP_ACCOUNTING
#define MAX_5_DIGIT 99999
static void
print_tu_heap_statistics(thread_units_t *tu, file_t logfile, const char *prefix)
{
int i;
size_t total = 0, cur = 0;
LOG(logfile, LOG_HEAP|LOG_STATS, 1, "%s heap breakdown:\n", prefix);
for (i = 0; i < ACCT_LAST; i++) {
/* print out cur since this is done periodically, not just at end */
LOG(logfile, LOG_HEAP|LOG_STATS, 1,
"%12s: cur=%5"SZFC"K, max=%5"SZFC"K, #=%7d, 1=",
whichheap_name[i], tu->acct.cur_usage[i]/1024,
tu->acct.max_usage[i]/1024, tu->acct.num_alloc[i]);
if (tu->acct.max_single[i] <= MAX_5_DIGIT)
LOG(logfile, LOG_HEAP|LOG_STATS, 1, "%5"SZFC, tu->acct.max_single[i]);
else {
LOG(logfile, LOG_HEAP|LOG_STATS, 1, "%4"SZFC"K",
tu->acct.max_single[i]/1024);
}
LOG(logfile, LOG_HEAP|LOG_STATS, 1,
", new=%5"SZFC"K, re=%5"SZFC"K\n",
tu->acct.alloc_new[i]/1024, tu->acct.alloc_reuse[i]/1024);
total += tu->acct.max_usage[i];
cur += tu->acct.cur_usage[i];
}
LOG(logfile, LOG_HEAP|LOG_STATS, 1,
"Total cur usage: %6"SZFC" KB\n", cur/1024);
LOG(logfile, LOG_HEAP|LOG_STATS, 1,
"Total max (not nec. all used simult.): %6"SZFC" KB\n", total/1024);
}
void
print_heap_statistics()
{
/* just do cur thread, don't try to walk all threads */
dcontext_t *dcontext = get_thread_private_dcontext();
DOSTATS({
uint i;
LOG(GLOBAL, LOG_STATS, 1, "Heap bucket usage counts and wasted memory:\n");
for (i=0; i<BLOCK_TYPES; i++) {
LOG(GLOBAL, LOG_STATS|LOG_HEAP, 1,
"%2d %3d count=%9u peak_count=%9u peak_wasted=%9u peak_align=%9u\n",
i, BLOCK_SIZES[i], block_total_count[i], block_peak_count[i],
block_peak_wasted[i], block_peak_align_pad[i]);
}
});
if (dcontext != NULL) {
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
if (th != NULL) { /* may not be initialized yet */
print_tu_heap_statistics(th->local_heap, THREAD, "Thread");
if (DYNAMO_OPTION(enable_reset)) {
ASSERT(th->nonpersistent_heap != NULL);
print_tu_heap_statistics(th->nonpersistent_heap, THREAD,
"Thread non-persistent");
}
}
}
if (DYNAMO_OPTION(enable_reset)) {
print_tu_heap_statistics(&heapmgt->global_nonpersistent_units, GLOBAL,
"Non-persistent global units");
}
print_tu_heap_statistics(&global_racy_units, GLOBAL, "Racy Up-to-date Process");
print_tu_heap_statistics(&heapmgt->global_units, GLOBAL,
"Updated-at-end Process (max is total of maxes)");
}
static void
add_heapacct_to_global_stats(heap_acct_t *acct)
{
/* add this thread's stats to the accurate (non-racy) global stats
* FIXME: this gives a nice in-one-place total, but loses the
* global-heap-only stats -- perhaps should add a total_units stats
* to capture total and leave global alone here?
*/
uint i;
acquire_recursive_lock(&global_alloc_lock);
for (i = 0; i < ACCT_LAST; i++) {
heapmgt->global_units.acct.alloc_reuse[i] += acct->alloc_reuse[i];
heapmgt->global_units.acct.alloc_new[i] += acct->alloc_new[i];
heapmgt->global_units.acct.cur_usage[i] += acct->cur_usage[i];
/* FIXME: these maxes are now not simultaneous max but sum-of-maxes */
heapmgt->global_units.acct.max_usage[i] += acct->max_usage[i];
heapmgt->global_units.acct.max_single[i] += acct->max_single[i];
heapmgt->global_units.acct.num_alloc[i] += acct->num_alloc[i];
}
release_recursive_lock(&global_alloc_lock);
}
#endif
/* dcontext only used for debugging */
static void
threadunits_exit(thread_units_t *tu, dcontext_t *dcontext)
{
heap_unit_t *u, *next_u;
#ifdef DEBUG
size_t total_heap_used = 0;
# ifdef HEAP_ACCOUNTING
int j;
# endif
#endif
#ifdef DEBUG_MEMORY
/* verify and clear (for later asserts) the free list */
uint i;
for (i = 0; i < BLOCK_TYPES; i++) {
heap_pc p, next_p;
for (p = tu->free_list[i]; p != NULL; p = next_p) {
next_p = *(heap_pc *)p;
/* clear the pointer to the next free for later asserts */
*(heap_pc *)p = (heap_pc) HEAP_UNALLOCATED_PTR_UINT;
DOCHECK(CHKLVL_MEMFILL, {
if (i < BLOCK_TYPES-1) {
CLIENT_ASSERT(is_region_memset_to_char(p, BLOCK_SIZES[i],
HEAP_UNALLOCATED_BYTE),
"memory corruption detected");
} else {
/* variable sized blocks */
CLIENT_ASSERT(is_region_memset_to_char(p, VARIABLE_SIZE(p),
HEAP_UNALLOCATED_BYTE),
"memory corruption detected");
/* clear the header for later asserts */
MEMSET_HEADER(p, HEAP_UNALLOCATED);
}
});
}
tu->free_list[i] = NULL;
}
#endif
u = tu->top_unit;
while (u != NULL) {
DOLOG(1, LOG_HEAP|LOG_STATS, {
size_t num_used = u->cur_pc - u->start_pc;
total_heap_used += num_used;
LOG(THREAD,
LOG_HEAP|LOG_STATS, 1,
"Heap unit %d @"PFX"-"PFX" [-"PFX"] ("SZFMT" [/"SZFMT"] KB): used "
SZFMT" KB\n",
u->id, u, UNIT_COMMIT_END(u),
UNIT_RESERVED_END(u), (UNIT_COMMIT_SIZE(u))/1024,
(UNIT_RESERVED_SIZE(u))/1024, num_used/1024);
});
next_u = u->next_local;
heap_free_unit(u, dcontext);
u = next_u;
}
LOG(THREAD, LOG_HEAP|LOG_STATS, 1,
"\tTotal heap used: "SZFMT" KB\n", total_heap_used/1024);
#if defined(DEBUG) && defined(HEAP_ACCOUNTING)
/* FIXME: separate scopes - in smaller functions for each of DEBUG_MEMORY and HEAP_ACCOUNTING */
for (j = 0; j < ACCT_LAST; j++) {
size_t usage = tu->acct.cur_usage[j];
if (usage > 0) {
LOG(THREAD, LOG_HEAP|LOG_STATS, 1,
"WARNING: %s "SZFMT" bytes not freed!\n",
whichheap_name[j], tu->acct.cur_usage[j]);
# ifdef HOT_PATCHING_INTERFACE /* known leaks for case 9593 */
if (DYNAMO_OPTION(hotp_only) &&
((j == ACCT_SPECIAL && usage == (size_t)hotp_only_tramp_bytes_leaked) ||
/* +4 is for the allocation's header; internal to heap mgt. */
(j == ACCT_MEM_MGT &&
usage == (size_t)(get_special_heap_header_size() + 4) &&
hotp_only_tramp_bytes_leaked > 0)))
continue;
# endif
if (j != ACCT_TOMBSTONE /* known leak */ &&
/* i#157: private loader => system lib allocs come here =>
* they don't always clean up
*/
j != ACCT_LIBDUP &&
INTERNAL_OPTION(heap_accounting_assert)) {
SYSLOG_INTERNAL_ERROR("memory leak: %s "SZFMT" bytes not freed",
whichheap_name[j], tu->acct.cur_usage[j]);
/* Don't assert when client does premature exit as it's
* hard for Extension libs, etc. to clean up in such situations:
*/
CLIENT_ASSERT(IF_CLIENT_INTERFACE(client_requested_exit ||) false,
"memory leak detected");
}
}
}
if (tu != &heapmgt->global_units)
add_heapacct_to_global_stats(&tu->acct);
DOLOG(1, LOG_HEAP|LOG_STATS, {
print_tu_heap_statistics(tu, THREAD,
dcontext == GLOBAL_DCONTEXT ? "Process" : "Thread");
});
#endif /* defined(DEBUG) && defined(HEAP_ACCOUNTING) */
}
void
heap_thread_reset_init(dcontext_t *dcontext)
{
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
if (DYNAMO_OPTION(enable_reset)) {
ASSERT(th->nonpersistent_heap != NULL);
threadunits_init(dcontext, th->nonpersistent_heap, HEAP_UNIT_MIN_SIZE);
}
}
void
heap_thread_init(dcontext_t *dcontext)
{
thread_heap_t *th = (thread_heap_t *)
global_heap_alloc(sizeof(thread_heap_t) HEAPACCT(ACCT_MEM_MGT));
dcontext->heap_field = (void *) th;
th->local_heap = (thread_units_t *) global_heap_alloc(sizeof(thread_units_t)
HEAPACCT(ACCT_MEM_MGT));
threadunits_init(dcontext, th->local_heap, HEAP_UNIT_MIN_SIZE);
if (DYNAMO_OPTION(enable_reset)) {
th->nonpersistent_heap = (thread_units_t *)
global_heap_alloc(sizeof(thread_units_t) HEAPACCT(ACCT_MEM_MGT));
} else
th->nonpersistent_heap = NULL;
heap_thread_reset_init(dcontext);
}
void
heap_thread_reset_free(dcontext_t *dcontext)
{
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
if (DYNAMO_OPTION(enable_reset)) {
ASSERT(th->nonpersistent_heap != NULL);
/* FIXME: free directly rather than sending to dead list for
* heap_reset_free() to free!
* FIXME: for reset, don't free last unit so don't have to
* recreate in reset_init()
*/
threadunits_exit(th->nonpersistent_heap, dcontext);
}
}
void
heap_thread_exit(dcontext_t *dcontext)
{
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
threadunits_exit(th->local_heap, dcontext);
heap_thread_reset_free(dcontext);
global_heap_free(th->local_heap, sizeof(thread_units_t) HEAPACCT(ACCT_MEM_MGT));
if (DYNAMO_OPTION(enable_reset)) {
ASSERT(th->nonpersistent_heap != NULL);
global_heap_free(th->nonpersistent_heap, sizeof(thread_units_t)
HEAPACCT(ACCT_MEM_MGT));
}
global_heap_free(th, sizeof(thread_heap_t) HEAPACCT(ACCT_MEM_MGT));
}
#if defined(DEBUG_MEMORY) && defined(DEBUG)
void
print_free_list(thread_units_t *tu, int i)
{
void *p;
int len = 0;
dcontext_t *dcontext = tu->dcontext;
LOG(THREAD, LOG_HEAP, 1,
"Free list for size %d (== %d bytes):\n", i, BLOCK_SIZES[i]);
p = (void *) tu->free_list[i];
while (p != NULL) {
LOG(THREAD, LOG_HEAP, 1, "\tp = "PFX"\n", p);
len++;
p = *((char **)p);
}
LOG(THREAD, LOG_HEAP, 1, "Total length is %d\n", len);
}
#endif
/* Used for both heap_unit_t and special_heap_unit_t.
* Returns the amount it increased the unit by, so caller should increment
* end_pc.
* Both end_pc and reserved_end_pc are assumed to be open-ended!
*/
static size_t
common_heap_extend_commitment(heap_pc cur_pc, heap_pc end_pc, heap_pc reserved_end_pc,
size_t size_need, uint prot)
{
if (end_pc < reserved_end_pc &&
!POINTER_OVERFLOW_ON_ADD(cur_pc, size_need)) {
/* extend commitment if have more reserved */
size_t commit_size = DYNAMO_OPTION(heap_commit_increment);
/* simpler to just not support taking very last page in address space */
if (POINTER_OVERFLOW_ON_ADD(end_pc, commit_size))
return 0;
if (cur_pc + size_need > end_pc + commit_size) {
commit_size = ALIGN_FORWARD(cur_pc + size_need - (ptr_uint_t)end_pc,
PAGE_SIZE);
}
if (end_pc + commit_size > reserved_end_pc ||
POINTER_OVERFLOW_ON_ADD(end_pc, commit_size)/*overflow seen in PR 518644 */) {
/* commit anyway before caller moves on to new unit so that
* we keep an invariant that all units but the current one
* are fully committed, so our algorithm for looking at the end
* of prior units holds
*/
commit_size = reserved_end_pc - end_pc;
}
ASSERT(!POINTER_OVERFLOW_ON_ADD(end_pc, commit_size) &&
end_pc + commit_size <= reserved_end_pc);
extend_commitment(end_pc, commit_size, prot, false /* extension */);
#ifdef DEBUG_MEMORY
memset(end_pc, HEAP_UNALLOCATED_BYTE, commit_size);
#endif
/* caller should do end_pc += commit_size */
STATS_ADD_PEAK(heap_capacity, commit_size);
/* FIXME: heap sizes are not always page-aligned so stats will be off */
STATS_SUB(heap_reserved_only, commit_size);
ASSERT(end_pc <= reserved_end_pc);
return commit_size;
} else
return 0;
}
static void
heap_unit_extend_commitment(heap_unit_t *u, size_t size_need, uint prot)
{
u->end_pc +=
common_heap_extend_commitment(u->cur_pc, u->end_pc, u->reserved_end_pc,
size_need, prot);
}
/* allocate storage on the DR heap
* returns NULL iff caller needs to grab dynamo_vm_areas_lock() and retry
*/
static void*
common_heap_alloc(thread_units_t *tu, size_t size HEAPACCT(which_heap_t which))
{
heap_unit_t *u = tu->cur_unit;
heap_pc p = NULL;
int bucket = 0;
size_t alloc_size, aligned_size;
#if defined(DEBUG_MEMORY) && defined(DEBUG)
size_t check_alloc_size;
dcontext_t *dcontext = tu->dcontext;
/* DrMem i#999: private libs can be heap-intensive and our checks here
* can have a prohibitive perf cost!
*/
uint chklvl = CHKLVL_MEMFILL + (IF_HEAPACCT_ELSE(which == ACCT_LIBDUP ? 1 : 0, 0));
ASSERT_CURIOSITY(which != ACCT_TOMBSTONE &&
"Do you really need to use ACCT_TOMBSTONE? (potentially dangerous)");
#endif
ASSERT(size > 0); /* we don't want to pay check cost in release */
ASSERT(size < MAX_VALID_HEAP_ALLOCATION && "potential integer overflow");
/* we prefer to crash than having heap overflows */
if (size > MAX_VALID_HEAP_ALLOCATION) {
/* This routine can currently accommodate without integer
* overflows sizes up to UINT_MAX - sizeof(heap_unit_t), but
* INT_MAX should be more than enough.
*
* Caller will likely crash, but that is better than a heap
* overflow, where a crash would be the best we can hope for.
*/
return NULL;
}
/* NOTE - all of our buckets are sized to preserve alignment, so this can't change
* which bucket is used. */
aligned_size = ALIGN_FORWARD(size, HEAP_ALIGNMENT);
while (aligned_size > BLOCK_SIZES[bucket])
bucket++;
if (bucket == BLOCK_TYPES-1)
alloc_size = aligned_size + HEADER_SIZE;
else
alloc_size = BLOCK_SIZES[bucket];
ASSERT(size <= alloc_size);
#ifdef DEBUG_MEMORY
/* case 10292: use original calculated size for later check */
check_alloc_size = alloc_size;
#endif
if (alloc_size > MAXROOM) {
/* too big for normal unit, build a special unit just for this allocation */
/* don't need alloc_size or even aligned_size, just need size */
heap_unit_t *new_unit, *prev;
/* we page-align to avoid wasting space if unit gets reused later */
size_t unit_size = ALIGN_FORWARD(size + sizeof(heap_unit_t), PAGE_SIZE);
ASSERT(size < unit_size && "overflow");
if (!safe_to_allocate_or_free_heap_units()) {
/* circular dependence solution: we need to hold DR lock before
* global alloc lock -- so we back out, grab it, and then come back
*/
return NULL;
}
/* Can reuse a dead unit if large enough: we'll just not use any
* excess size until this is freed and put back on dead list.
* (Currently we don't put oversized units on dead list though.)
*/
new_unit = heap_create_unit(tu, unit_size, false/*can be reused*/);
/* we want to commit the whole alloc right away */
heap_unit_extend_commitment(new_unit, size, MEMPROT_READ|MEMPROT_WRITE);
prev = tu->top_unit;
alloc_size = size; /* should we include page-alignment? */
/* insert prior to cur unit (new unit will be full, so keep cur unit
* where it is)
*/
while (prev != u && prev->next_local != u) {
ASSERT(prev != NULL && prev->next_local != NULL);
prev = prev->next_local;
}
if (prev == u) {
ASSERT(prev == tu->top_unit);
tu->top_unit = new_unit;
} else
prev->next_local = new_unit;
new_unit->next_local = u;
#ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 3,
"\tCreating new oversized heap unit %d (%d [/%d] KB)\n",
new_unit->id, UNIT_COMMIT_SIZE(new_unit)/1024,
UNIT_RESERVED_SIZE(new_unit)/1024);
#endif
p = new_unit->start_pc;
new_unit->cur_pc += size;
ACCOUNT_FOR_ALLOC(alloc_new, tu, which, size, size); /* use alloc_size? */
goto done_allocating;
}
if (tu->free_list[bucket] != NULL) {
if (bucket == BLOCK_TYPES-1) {
/* variable-length blocks, try to find one big enough */
size_t sz;
heap_pc next = tu->free_list[bucket];
heap_pc prev;
do {
prev = p;
p = next;
/* aligned_size is written right _before_ next pointer */
sz = VARIABLE_SIZE(next);
next = *((heap_pc*)p);
} while (aligned_size > sz && next != NULL);
if (aligned_size <= sz) {
ASSERT(ALIGNED(next, HEAP_ALIGNMENT));
/* found one, extract from free list */
if (p == tu->free_list[bucket])
tu->free_list[bucket] = next;
else
*((heap_pc *)prev) = next;
#ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 2,
"Variable-size block: allocating "PFX" (%d bytes [%d aligned] in "
"%d block)\n", p, size, aligned_size, sz);
/* ensure memory we got from the free list is in a heap unit */
DOCHECK(CHKLVL_DEFAULT, { /* expensive check */
ASSERT(find_heap_unit(tu, p, sz) != NULL);
});
#endif
ASSERT(ALIGNED(sz, HEAP_ALIGNMENT));
alloc_size = sz + HEADER_SIZE;
ACCOUNT_FOR_ALLOC(alloc_reuse, tu, which, alloc_size, aligned_size);
} else {
/* no free block big enough available */
p = NULL;
}
} else {
/* fixed-length free block available */
p = tu->free_list[bucket];
tu->free_list[bucket] = *((heap_pc *)p);
ASSERT(ALIGNED(tu->free_list[bucket], HEAP_ALIGNMENT));
#ifdef DEBUG_MEMORY
/* ensure memory we got from the free list is in a heap unit */
DOCHECK(CHKLVL_DEFAULT, { /* expensive check */
ASSERT(find_heap_unit(tu, p, alloc_size) != NULL);
});
#endif
ACCOUNT_FOR_ALLOC(alloc_reuse, tu, which, alloc_size, aligned_size);
}
}
if (p == NULL) {
/* no free blocks, grab a new one */
/* FIXME: if no more heap but lots of larger blocks available,
* should use the larger blocks instead of failing! */
/* see if room for allocation size */
ASSERT(ALIGNED(u->cur_pc, HEAP_ALIGNMENT));
ASSERT(ALIGNED(alloc_size, HEAP_ALIGNMENT));
if (u->cur_pc + alloc_size > u->end_pc ||
POINTER_OVERFLOW_ON_ADD(u->cur_pc, alloc_size)/*xref PR 495961*/) {
/* We either have to extend the current unit or, failing that,
* allocate a new unit. */
if (!safe_to_allocate_or_free_heap_units()) {
/* circular dependence solution: we need to hold dynamo areas
* lock before global alloc lock in case we end up adding a new
* unit or we hit oom (which may free units) while extending
* the commmitment -- so we back out, grab it, and then come
* back. */
return NULL;
}
/* try to extend if possible */
heap_unit_extend_commitment(u, alloc_size, MEMPROT_READ|MEMPROT_WRITE);
/* check again after extending commit */
if (u->cur_pc + alloc_size > u->end_pc ||
POINTER_OVERFLOW_ON_ADD(u->cur_pc, alloc_size)/*xref PR 495961*/) {
/* no room, look for room at end of previous units
* FIXME: instead should put end of unit space on free list!
*/
heap_unit_t *prev = tu->top_unit;
while (1) {
/* make sure we do NOT steal space from oversized units,
* who though they may have extra space from alignment
* may be freed wholesale when primary alloc is freed
*/
if (UNITALLOC(prev) <= HEAP_UNIT_MAX_SIZE &&
!POINTER_OVERFLOW_ON_ADD(prev->cur_pc, alloc_size) &&
prev->cur_pc + alloc_size <= prev->end_pc) {
tu->cur_unit = prev;
u = prev;
break;
}
if (prev->next_local == NULL) {
/* no room anywhere, so create new unit
* double size of old unit (until hit max size)
*/
heap_unit_t *new_unit;
size_t unit_size;
unit_size = UNITALLOC(u) * 2;
while (unit_size < alloc_size + UNITOVERHEAD)
unit_size *= 2;
if (unit_size > HEAP_UNIT_MAX_SIZE)
unit_size = HEAP_UNIT_MAX_SIZE;
/* size for heap_create_unit doesn't include any guard
* pages */
ASSERT(unit_size > UNITOVERHEAD);
ASSERT(unit_size > (size_t) GUARD_PAGE_ADJUSTMENT);
unit_size -= GUARD_PAGE_ADJUSTMENT;
new_unit = heap_create_unit(tu, unit_size, false/*can reuse*/);
prev->next_local = new_unit;
#ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 2,
"\tCreating new heap unit %d (%d [/%d] KB)\n",
new_unit->id, UNIT_COMMIT_SIZE(new_unit)/1024,
UNIT_RESERVED_SIZE(new_unit)/1024);
#endif
/* use new unit for all future non-free-list allocations
* we'll try to use the free room at the end of the old unit(s)
* only when we next run out of room
*/
tu->cur_unit = new_unit;
u = new_unit;
/* may need to extend now if alloc_size is large */
heap_unit_extend_commitment(u, alloc_size,
MEMPROT_READ|MEMPROT_WRITE);
/* otherwise would have been bigger than MAXROOM */
ASSERT(alloc_size <= (ptr_uint_t) (u->end_pc - u->cur_pc));
break;
}
prev = prev->next_local;
}
}
}
p = u->cur_pc;
if (bucket == BLOCK_TYPES-1) {
/* we keep HEADER_SIZE bytes to store the size */
p += HEADER_SIZE;
VARIABLE_SIZE(p) = aligned_size;
}
u->cur_pc += alloc_size;
ACCOUNT_FOR_ALLOC(alloc_new, tu, which, alloc_size, aligned_size);
}
DOSTATS({
/* do this before done_allocating: want to ignore special-unit allocs */
ATOMIC_ADD(int, block_count[bucket], 1);
ATOMIC_ADD(int, block_total_count[bucket], 1);
/* FIXME: should atomically store inc-ed val in temp to avoid races w/ max */
ATOMIC_MAX(int, block_peak_count[bucket], block_count[bucket]);
ASSERT(CHECK_TRUNCATE_TYPE_uint(alloc_size - aligned_size));
ATOMIC_ADD(int, block_wasted[bucket], (int) (alloc_size - aligned_size));
/* FIXME: should atomically store val in temp to avoid races w/ max */
ATOMIC_MAX(int, block_peak_wasted[bucket], block_wasted[bucket]);
if (aligned_size > size) {
ASSERT(CHECK_TRUNCATE_TYPE_uint(aligned_size - size));
ATOMIC_ADD(int, block_align_pad[bucket], (int) (aligned_size - size));
/* FIXME: should atomically store val in temp to avoid races w/ max */
ATOMIC_MAX(int, block_peak_align_pad[bucket], block_align_pad[bucket]);
STATS_ADD_PEAK(heap_align, aligned_size - size);
LOG(GLOBAL, LOG_STATS, 5,
"alignment mismatch: %s ask %d, aligned is %d -> %d pad\n",
IF_HEAPACCT_ELSE(whichheap_name[which], ""),
size, aligned_size, aligned_size-size);
}
if (bucket == BLOCK_TYPES-1) {
STATS_ADD(heap_headers, HEADER_SIZE);
STATS_INC(heap_allocs_variable);
} else {
STATS_INC(heap_allocs_buckets);
if (alloc_size > aligned_size) {
STATS_ADD_PEAK(heap_bucket_pad, alloc_size - aligned_size);
LOG(GLOBAL, LOG_STATS, 5,
"bucket mismatch: %s ask (aligned) %d, got %d, -> %d\n",
IF_HEAPACCT_ELSE(whichheap_name[which], ""),
aligned_size, alloc_size, alloc_size-aligned_size);
}
}
});
done_allocating:
#ifdef DEBUG_MEMORY
if (bucket == BLOCK_TYPES-1 && check_alloc_size <= MAXROOM) {
/* verify is unallocated memory, skip possible free list next pointer */
DOCHECK(chklvl, {
CLIENT_ASSERT(is_region_memset_to_char
(p+sizeof(heap_pc *), (alloc_size-HEADER_SIZE)-sizeof(heap_pc *),
HEAP_UNALLOCATED_BYTE), "memory corruption detected");
});
LOG(THREAD, LOG_HEAP, 6,
"\nalloc var "PFX"-"PFX" %d bytes, ret "PFX"-"PFX" %d bytes\n",
p-HEADER_SIZE, p-HEADER_SIZE+alloc_size, alloc_size, p, p+size, size);
/* there can only be extra padding if we took off of the free list */
DOCHECK(chklvl, memset(p+size, HEAP_PAD_BYTE, (alloc_size-HEADER_SIZE)-size););
} else {
/* verify is unallocated memory, skip possible free list next pointer */
DOCHECK(chklvl, {
CLIENT_ASSERT(is_region_memset_to_char
(p+sizeof(heap_pc *), alloc_size-sizeof(heap_pc *),
HEAP_UNALLOCATED_BYTE), "memory corruption detected");
});
LOG(THREAD, LOG_HEAP, 6,
"\nalloc fix or oversize "PFX"-"PFX" %d bytes, ret "PFX"-"PFX" %d bytes\n",
p, p+alloc_size, alloc_size, p, p+size, size);
DOCHECK(chklvl, memset(p+size, HEAP_PAD_BYTE, alloc_size-size););
}
DOCHECK(chklvl, memset(p, HEAP_ALLOCATED_BYTE, size););
# ifdef HEAP_ACCOUNTING
LOG(THREAD, LOG_HEAP, 6, "\t%s\n", whichheap_name[which]);
# endif
#endif
return (void*)p;
}
/* allocate storage on the thread's private heap */
void*
heap_alloc(dcontext_t *dcontext, size_t size HEAPACCT(which_heap_t which))
{
thread_units_t *tu;
void *ret_val;
if (dcontext == GLOBAL_DCONTEXT)
return global_heap_alloc(size HEAPACCT(which));
tu = ((thread_heap_t *) dcontext->heap_field)->local_heap;
ret_val = common_heap_alloc(tu, size HEAPACCT(which));
ASSERT(ret_val != NULL);
return ret_val;
}
/* free heap storage
* returns false if caller needs to grab dynamo_vm_areas_lock() and retry
*/
static bool
common_heap_free(thread_units_t *tu, void *p_void, size_t size HEAPACCT(which_heap_t which))
{
int bucket = 0;
heap_pc p = (heap_pc) p_void;
#if defined(DEBUG) && (defined(DEBUG_MEMORY) || defined(HEAP_ACCOUNTING))
dcontext_t *dcontext = tu->dcontext;
/* DrMem i#999: private libs can be heap-intensive and our checks here
* can have a prohibitive perf cost!
* XXX: b/c of re-use we have to memset on free. Perhaps we should
* have a separate heap pool for private libs. But, the overhead
* from that final memset is small compared to what we've already
* saved, so maybe not worth it.
*/
uint chklvl = CHKLVL_MEMFILL + (IF_HEAPACCT_ELSE(which == ACCT_LIBDUP ? 1 : 0, 0));
#endif
size_t alloc_size, aligned_size = ALIGN_FORWARD(size, HEAP_ALIGNMENT);
ASSERT(size > 0); /* we don't want to pay check cost in release */
ASSERT(p != NULL);
#ifdef DEBUG_MEMORY
/* FIXME i#417: This curiosity assertion is trying to make sure we don't
* perform a double free, but it can fire if we ever free a data structure
* that has the 0xcdcdcdcd bitpattern in the first or last 4 bytes. This
* has happened a few times:
*
* - case 8802: App's eax is 0xcdcdcdcd (from an app dbg memset) and we have
* dcontext->allocated_start==dcontext.
* - i#417: On Linux x64 we get rax == 0xcdcdcdcd from a memset, and
* opnd_create_reg() only updates part of the register before returning by
* value in RAX:RDX. We initialize to zero in debug mode to work around
* this.
* - i#540: On Win7 x64 we see this assert when running the TSan tests in
* NegativeTests.WindowsRegisterWaitForSingleObjectTest.
*
* For now, we've downgraded this to a curiosity, but if it fires too much
* in the future we should maintain a separate data structure in debug mode
* to perform this check. We accept objects that start with 0xcdcdcdcd so
* long as the second four bytes are not also 0xcdcdcdcd.
*/
DOCHECK(chklvl, {
ASSERT_CURIOSITY(
(*(uint *)p != HEAP_UNALLOCATED_UINT ||
(size >= 2*sizeof(uint) && *(((uint *)p)+1) != HEAP_UNALLOCATED_UINT)) &&
*(uint *)(p+size-sizeof(int)) != HEAP_UNALLOCATED_UINT &&
"attempting to free memory containing HEAP_UNALLOCATED pattern, "
"possible double free!");
});
#endif
while (aligned_size > BLOCK_SIZES[bucket])
bucket++;
if (bucket == BLOCK_TYPES-1)
alloc_size = aligned_size + HEADER_SIZE;
else
alloc_size = BLOCK_SIZES[bucket];
if (alloc_size > MAXROOM) {
/* we must have used a special unit just for this allocation */
heap_unit_t *u = tu->top_unit, *prev = NULL;
#ifdef DEBUG_MEMORY
/* ensure we are freeing memory in a proper unit */
DOCHECK(CHKLVL_DEFAULT, { /* expensive check */
ASSERT(find_heap_unit(tu, p, size) != NULL);
});
#endif
if (!safe_to_allocate_or_free_heap_units()) {
/* circular dependence solution: we need to hold DR lock before
* global alloc lock -- so we back out, grab it, and then come back
*/
return false;
}
while (u != NULL && u->start_pc != p) {
prev = u;
u = u->next_local;
}
ASSERT(u != NULL);
/* remove this unit from this thread's list, move to dead list
* for future use -- no problems will be caused by it being
* larger than normal
*/
if (prev == NULL)
tu->top_unit = u->next_local;
else
prev->next_local = u->next_local;
/* just retire the unit # */
#ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 3, "\tFreeing oversized heap unit %d (%d KB)\n",
u->id, size/1024);
/* go ahead and set unallocated, even though we are just going to free
* the unit, is needed for an assert in heap_free_unit anyways */
DOCHECK(CHKLVL_MEMFILL, memset(p, HEAP_UNALLOCATED_BYTE, size););
#endif
ASSERT(size <= UNITROOM(u));
heap_free_unit(u, tu->dcontext);
ACCOUNT_FOR_FREE(tu, which, size);
return true;
} else if (bucket == BLOCK_TYPES-1) {
ASSERT(GET_VARIABLE_ALLOCATION_SIZE(p) >= alloc_size);
alloc_size = GET_VARIABLE_ALLOCATION_SIZE(p);
ASSERT(alloc_size - HEADER_SIZE >= aligned_size);
}
#if defined(DEBUG) || defined(DEBUG_MEMORY) || defined(HEAP_ACCOUNTING)
if (bucket == BLOCK_TYPES-1) {
# ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 6,
"\nfree var "PFX"-"PFX" %d bytes, asked "PFX"-"PFX" %d bytes\n",
p-HEADER_SIZE, p-HEADER_SIZE+alloc_size, alloc_size, p, p+size, size);
ASSERT_MESSAGE(chklvl, "heap overflow",
is_region_memset_to_char(p+size, (alloc_size-HEADER_SIZE)-size,
HEAP_PAD_BYTE));
/* ensure we are freeing memory in a proper unit */
DOCHECK(CHKLVL_DEFAULT, { /* expensive check */
ASSERT(find_heap_unit(tu, p, alloc_size - HEADER_SIZE) != NULL);
});
/* set used and padding memory back to unallocated */
DOCHECK(CHKLVL_MEMFILL, memset(p, HEAP_UNALLOCATED_BYTE, alloc_size-HEADER_SIZE););
# endif
STATS_SUB(heap_headers, HEADER_SIZE);
} else {
# ifdef DEBUG_MEMORY
LOG(THREAD, LOG_HEAP, 6,
"\nfree fix "PFX"-"PFX" %d bytes, asked "PFX"-"PFX" %d bytes\n",
p, p+alloc_size, alloc_size, p, p+size, size);
ASSERT_MESSAGE(chklvl, "heap overflow",
is_region_memset_to_char(p+size, alloc_size-size, HEAP_PAD_BYTE));
/* ensure we are freeing memory in a proper unit */
DOCHECK(CHKLVL_DEFAULT, { /* expensive check */
ASSERT(find_heap_unit(tu, p, alloc_size) != NULL);
});
/* set used and padding memory back to unallocated */
DOCHECK(CHKLVL_MEMFILL, memset(p, HEAP_UNALLOCATED_BYTE, alloc_size););
# endif
STATS_SUB(heap_bucket_pad, (alloc_size - aligned_size));
}
STATS_SUB(heap_align, (aligned_size - size));
DOSTATS({
ATOMIC_ADD(int, block_count[bucket], -1);
ATOMIC_ADD(int, block_wasted[bucket], -(int)(alloc_size - aligned_size));
ATOMIC_ADD(int, block_align_pad[bucket], -(int)(aligned_size - size));
});
# ifdef HEAP_ACCOUNTING
LOG(THREAD, LOG_HEAP, 6, "\t%s\n", whichheap_name[which]);
ACCOUNT_FOR_FREE(tu, which, alloc_size);
# endif
#endif
/* write next pointer */
*((heap_pc*)p) = tu->free_list[bucket];
ASSERT(ALIGNED(tu->free_list[bucket], HEAP_ALIGNMENT));
tu->free_list[bucket] = p;
ASSERT(ALIGNED(tu->free_list[bucket], HEAP_ALIGNMENT));
return true;
}
/* free heap storage */
void
heap_free(dcontext_t *dcontext, void *p, size_t size HEAPACCT(which_heap_t which))
{
thread_units_t *tu;
DEBUG_DECLARE(bool ok;)
if (dcontext == GLOBAL_DCONTEXT) {
global_heap_free(p, size HEAPACCT(which));
return;
}
tu = ((thread_heap_t *) dcontext->heap_field)->local_heap;
DEBUG_DECLARE(ok = ) common_heap_free(tu, p, size HEAPACCT(which));
ASSERT(ok);
}
bool local_heap_protected(dcontext_t *dcontext)
{
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
return (!th->local_heap->writable ||
(th->nonpersistent_heap != NULL && !th->nonpersistent_heap->writable));
}
static inline void
protect_local_units_helper(heap_unit_t *u, bool writable)
{
/* win32 does not allow single protection change call on units that
* were allocated with separate calls so we don't try to combine
* adjacent units here
*/
while (u != NULL) {
change_protection(UNIT_ALLOC_START(u), UNIT_COMMIT_SIZE(u), writable);
u = u->next_local;
}
}
static void
protect_threadunits(thread_units_t *tu, bool writable)
{
ASSERT(TEST(SELFPROT_LOCAL, dynamo_options.protect_mask));
if (tu->writable == writable)
return;
protect_local_units_helper(tu->top_unit, writable);
tu->writable = writable;
}
void
protect_local_heap(dcontext_t *dcontext, bool writable)
{
thread_heap_t *th = (thread_heap_t *) dcontext->heap_field;
protect_threadunits(th->local_heap, writable);
if (DYNAMO_OPTION(enable_reset))
protect_threadunits(th->nonpersistent_heap, writable);
}
/* assumption: vmm_heap_alloc only gets called for HeapUnits themselves, which
* are protected by us here, so ignore os heap
*/
void
protect_global_heap(bool writable)
{
ASSERT(TEST(SELFPROT_GLOBAL, dynamo_options.protect_mask));
acquire_recursive_lock(&global_alloc_lock);
if (heapmgt->global_heap_writable == writable) {
release_recursive_lock(&global_alloc_lock);
return;
}
/* win32 does not allow single protection change call on units that
* were allocated with separate calls so we don't try to combine
* adjacent units here
* FIXME: That may no longer be true for our virtual memory manager that
* will in fact be allocated as a single unit. It is only in case
* we have run out of that initial allocation that we may have to
* keep a separate list of allocations.
*/
if (!writable) {
ASSERT(heapmgt->global_heap_writable);
heapmgt->global_heap_writable = writable;
}
protect_local_units_helper(heapmgt->global_units.top_unit, writable);
if (DYNAMO_OPTION(enable_reset)) {
protect_local_units_helper(heapmgt->global_nonpersistent_units.top_unit,
writable);
}
if (writable) {
ASSERT(!heapmgt->global_heap_writable);
heapmgt->global_heap_writable = writable;
}
release_recursive_lock(&global_alloc_lock);
}
/* FIXME: share some code...right now these are identical to protected
* versions except the unit used
*/
void *
global_unprotected_heap_alloc(size_t size HEAPACCT(which_heap_t which))
{
void *p = common_global_heap_alloc(&heapmgt->global_unprotected_units,
size HEAPACCT(which));
ASSERT(p != NULL);
LOG(GLOBAL, LOG_HEAP, 6, "\nglobal unprotected alloc: "PFX" (%d bytes)\n", p, size);
return p;
}
void
global_unprotected_heap_free(void *p, size_t size HEAPACCT(which_heap_t which))
{
common_global_heap_free(&heapmgt->global_unprotected_units,
p, size HEAPACCT(which));
LOG(GLOBAL, LOG_HEAP, 6, "\nglobal unprotected free: "PFX" (%d bytes)\n", p, size);
}
void *
nonpersistent_heap_alloc(dcontext_t *dcontext, size_t size HEAPACCT(which_heap_t which))
{
void *p;
if (DYNAMO_OPTION(enable_reset)) {
if (dcontext == GLOBAL_DCONTEXT) {
p = common_global_heap_alloc(&heapmgt->global_nonpersistent_units,
size HEAPACCT(which));
LOG(GLOBAL, LOG_HEAP, 6,
"\nglobal nonpersistent alloc: "PFX" (%d bytes)\n", p, size);
} else {
thread_units_t *nph = ((thread_heap_t *) dcontext->heap_field)->nonpersistent_heap;
p = common_heap_alloc(nph, size HEAPACCT(which));
}
} else {
p = heap_alloc(dcontext, size HEAPACCT(which));
}
ASSERT(p != NULL);
return p;
}
void
nonpersistent_heap_free(dcontext_t *dcontext, void *p, size_t size
HEAPACCT(which_heap_t which))
{
if (DYNAMO_OPTION(enable_reset)) {
if (dcontext == GLOBAL_DCONTEXT) {
common_global_heap_free(&heapmgt->global_nonpersistent_units,
p, size HEAPACCT(which));
LOG(GLOBAL, LOG_HEAP, 6,
"\nglobal nonpersistent free: "PFX" (%d bytes)\n", p, size);
} else {
thread_units_t *nph = ((thread_heap_t *) dcontext->heap_field)->nonpersistent_heap;
DEBUG_DECLARE(bool ok =) common_heap_free(nph, p, size HEAPACCT(which));
ASSERT(ok);
}
} else {
heap_free(dcontext, p, size HEAPACCT(which));
}
}
/****************************************************************************
* SPECIAL SINGLE-ALLOC-SIZE HEAP SERVICE
*/
/* Assumptions:
* All allocations are of a single block size
* If use_lock is false, no synchronization is needed or even safe
*/
/* We use our own unit struct to give us flexibility.
* 1) We don't always allocate the header inline.
* 2) We are sometimes executed from and so need pc prof support.
* 3) We don't need all the fields of heap_unit_t.
*/
typedef struct _special_heap_unit_t {
heap_pc alloc_pc; /* start of allocation region */
heap_pc start_pc; /* first address we'll give out for storage */
heap_pc end_pc; /* open-ended address of heap storage */
heap_pc cur_pc; /* current end (open) of allocated storage */
heap_pc reserved_end_pc; /* (open) end of reserved (not nec committed) memory */
#ifdef WINDOWS_PC_SAMPLE
profile_t *profile;
#endif
#ifdef DEBUG
int id; /* # of this unit */
#endif
struct _special_heap_unit_t *next;
} special_heap_unit_t;
#define SPECIAL_UNIT_COMMIT_SIZE(u) ((u)->end_pc - (u)->alloc_pc)
#define SPECIAL_UNIT_RESERVED_SIZE(u) ((u)->reserved_end_pc - (u)->alloc_pc)
#define SPECIAL_UNIT_HEADER_INLINE(u) ((u)->alloc_pc != (u)->start_pc)
/* the cfree list stores a next ptr and a count */
typedef struct _cfree_header {
struct _cfree_header *next_cfree;
uint count;
} cfree_header_t;
typedef struct _special_units_t {
special_heap_unit_t *top_unit; /* start of linked list of heap units */
special_heap_unit_t *cur_unit; /* current unit in heap list */
uint block_size; /* all blocks are this size */
heap_pc free_list;
cfree_header_t *cfree_list;
#ifdef DEBUG
int num_units; /* total # of heap units */
#endif
bool writable:1; /* remember state of heap protection */
bool executable:1;
/* if use_lock is false, grabbing _any_ lock may be hazardous!
* (this isn't just an optimization, it's for correctness)
*/
bool use_lock:1;
bool in_iterator:1;
bool persistent:1;
mutex_t lock;
/* Yet another feature added: pclookup, but across multiple heaps,
* so it's via a passed-in vector and passed-in data
*/
vm_area_vector_t *heap_areas;
void *lookup_retval;
#ifdef WINDOWS_PC_SAMPLE
struct _special_units_t *next;
#endif
#ifdef HEAP_ACCOUNTING
/* we only need one bucket for SpecialHeap but to re-use code we waste space */
heap_acct_t acct;
#endif
} special_units_t;
#if defined(WINDOWS_PC_SAMPLE) && !defined(DEBUG)
/* For fast exit path we need a quick way to walk all the units */
DECLARE_CXTSWPROT_VAR(static mutex_t special_units_list_lock,
INIT_LOCK_FREE(special_units_list_lock));
/* This is only used for profiling so we don't bother to protect it */
DECLARE_CXTSWPROT_VAR(static special_units_t *special_units_list, NULL);
#endif
#if defined(DEBUG) && defined(HEAP_ACCOUNTING) && defined(HOT_PATCHING_INTERFACE)
/* To get around the problem of the special_units_t "module" being defined after
* the heap module in the same file. Part of fix for case 9593 that required
* leaking memory.
*/
static int
get_special_heap_header_size(void)
{
return sizeof(special_units_t);
}
#endif
#ifdef WINDOWS_PC_SAMPLE
static inline bool special_heap_profile_enabled()
{
return (dynamo_options.profile_pcs &&
dynamo_options.prof_pcs_stubs >= 2 &&
dynamo_options.prof_pcs_stubs <= 32);
}
#endif
static inline uint
get_prot(special_units_t *su)
{
return (su->executable ?
MEMPROT_READ|MEMPROT_WRITE|MEMPROT_EXEC :
MEMPROT_READ|MEMPROT_WRITE);
}
static void
special_unit_extend_commitment(special_heap_unit_t *u, size_t size_need, uint prot)
{
u->end_pc +=
common_heap_extend_commitment(u->cur_pc, u->end_pc, u->reserved_end_pc,
size_need, prot);
}
/* If pc is NULL, allocates memory and stores the header inside it;
* If pc is non-NULL, allocates separate memory for the header, and
* uses pc for the heap region (assuming size is fully committed).
* unit_full only applies to the non-NULL case, indicating whether
* to continue to allocate from this unit.
*/
static special_heap_unit_t *
special_heap_create_unit(special_units_t *su, byte *pc, size_t size, bool unit_full)
{
special_heap_unit_t *u;
size_t commit_size;
uint prot = get_prot(su);
ASSERT_OWN_MUTEX(su->use_lock, &su->lock);
if (pc != NULL) {
u = HEAP_TYPE_ALLOC(GLOBAL_DCONTEXT, special_heap_unit_t, ACCT_MEM_MGT, PROTECTED);
ASSERT(u != NULL);
u->start_pc = pc;
u->alloc_pc = pc;
commit_size = size;
/* caller should arrange alignment */
ASSERT(!IS_POWER_OF_2(su->block_size) || ALIGNED(u->start_pc, su->block_size));
} else {
commit_size = DYNAMO_OPTION(heap_commit_increment);
ASSERT(commit_size <= size);
/* create new unit */
/* Since vmm lock, dynamo_vm_areas lock, all_memory_areas lock (on
* linux), etc. will be acquired, and presumably !su->use_lock means
* user can't handle ANY lock being acquired, we assert here: xref PR
* 596768. In release build, we try to acqure the memory anyway. I'm
* worried about pcprofile: can only fit ~1K in one unit and so will
* easily run out...should it allocate additional units up front?
* => PR 596808.
*/
ASSERT(su->top_unit == NULL/*init*/ || su->use_lock);
u = (special_heap_unit_t *) get_guarded_real_memory(size, commit_size, prot,
true, true
_IF_DEBUG("special_heap"));
ASSERT(u != NULL);
u->alloc_pc = (heap_pc) u;
/* u is kept at top of unit itself, so displace start pc */
u->start_pc = (heap_pc) (((ptr_uint_t)u) + sizeof(special_heap_unit_t));
/* some users expect alignment; not much of a space loss for those who don't */
if (IS_POWER_OF_2(su->block_size)) {
STATS_ADD(heap_special_align,
ALIGN_FORWARD(u->start_pc, su->block_size) -
(ptr_uint_t)u->start_pc);
u->start_pc = (heap_pc) ALIGN_FORWARD(u->start_pc, su->block_size);
}
}
u->end_pc = u->alloc_pc + commit_size;
u->reserved_end_pc = u->alloc_pc + size;
if (pc != NULL && unit_full) {
ASSERT(u->reserved_end_pc == u->end_pc);
u->cur_pc = u->end_pc;
} else
u->cur_pc = u->start_pc;
u->next = NULL;
DODEBUG({
u->id = su->num_units;
su->num_units++;
});
#ifdef WINDOWS_PC_SAMPLE
if (special_heap_profile_enabled()) {
u->profile = create_profile((app_pc)PAGE_START(u->start_pc), u->reserved_end_pc,
dynamo_options.prof_pcs_stubs, NULL);
start_profile(u->profile);
} else
u->profile = NULL;
#endif
/* N.B.: if STATS macros ever change to grab a mutex, we could deadlock
* if !su->use_lock!
*/
STATS_ADD_PEAK(heap_capacity, commit_size);
STATS_ADD_PEAK(heap_special_capacity, commit_size);
STATS_ADD_PEAK(heap_special_units, 1);
STATS_ADD_PEAK(heap_reserved_only, (u->reserved_end_pc - u->end_pc));
if (su->heap_areas != NULL) {
vmvector_add(su->heap_areas, u->alloc_pc, u->reserved_end_pc, su->lookup_retval);
}
#ifdef DEBUG_MEMORY
/* Don't clobber already-allocated memory */
DOCHECK(CHKLVL_MEMFILL, {
if (pc == NULL)
memset(u->start_pc, HEAP_UNALLOCATED_BYTE, u->end_pc - u->start_pc);
});
#endif
return u;
}
/* caller must store the special_units_t *, which is opaque */
static void *
special_heap_init_internal(uint block_size, bool use_lock, bool executable,
bool persistent, vm_area_vector_t *vector, void *vector_data,
byte *heap_region, size_t heap_size, bool unit_full)
{
special_units_t *su;
size_t unit_size = (block_size * 16 > HEAP_UNIT_MIN_SIZE) ?
(block_size * 16) : HEAP_UNIT_MIN_SIZE;
/* Whether using 16K or 64K vmm blocks, HEAP_UNIT_MIN_SIZE of 32K wastes
* space, and our main uses (stubs, whether global or coarse, and signal
* pending queue) don't need a lot of space, so shrinking.
* This tuning is a little fragile (just like for regular heap units and
* fcache units) so be careful when changing default parameters.
*/
unit_size = (size_t) ALIGN_FORWARD(unit_size, PAGE_SIZE);
ASSERT(unit_size > (size_t) GUARD_PAGE_ADJUSTMENT);
unit_size -= GUARD_PAGE_ADJUSTMENT;
su = (special_units_t *)
(persistent ? global_heap_alloc(sizeof(special_units_t) HEAPACCT(ACCT_MEM_MGT)) :
nonpersistent_heap_alloc(GLOBAL_DCONTEXT, sizeof(special_units_t)
HEAPACCT(ACCT_MEM_MGT)));
memset(su, 0, sizeof(*su));
ASSERT(block_size >= sizeof(heap_pc *) && "need room for free list ptrs");
ASSERT(block_size >= sizeof(heap_pc *) + sizeof(uint) &&
"need room for cfree list ptrs");
su->block_size = block_size;
su->executable = executable;
su->persistent = persistent;
su->writable = true;
su->free_list = NULL;
su->cfree_list = NULL;
DODEBUG({
su->num_units = 0;
});
ASSERT((vector == NULL) == (vector_data == NULL));
su->heap_areas = vector;
su->lookup_retval = vector_data;
su->in_iterator = false;
if (use_lock)
ASSIGN_INIT_LOCK_FREE(su->lock, special_heap_lock);
/* For persistent cache loading we hold executable_areas lock and so
* cannot acquire special_heap_lock -- so we do not acquire
* for the initial unit creation, which is safe since su is still
* private to this routine.
*/
su->use_lock = false; /* we set to real value below */
su->top_unit = special_heap_create_unit(su, heap_region,
heap_size == 0 ? unit_size : heap_size,
unit_full);
su->use_lock = use_lock;
#ifdef HEAP_ACCOUNTING
memset(&su->acct, 0, sizeof(su->acct));
#endif
su->cur_unit = su->top_unit;
#if defined(WINDOWS_PC_SAMPLE) && !defined(DEBUG)
if (special_heap_profile_enabled()) {
/* Add to the global master list, which requires a lock */
mutex_lock(&special_units_list_lock);
su->next = special_units_list;
special_units_list = su;
mutex_unlock(&special_units_list_lock);
}
#endif
return su;
}
/* Typical usage */
void *
special_heap_init(uint block_size, bool use_lock, bool executable, bool persistent)
{
return special_heap_init_internal(block_size, use_lock, executable,
persistent, NULL, NULL, NULL, 0, false);
}
/* Special heap w/ a vector for lookups. Also supports a pre-created heap region
* (heap_region, heap_region+heap_size) whose fullness is unit_full. */
void *
special_heap_pclookup_init(uint block_size, bool use_lock, bool executable,
bool persistent, vm_area_vector_t *vector, void *vector_data,
byte *heap_region, size_t heap_size, bool unit_full)
{
return special_heap_init_internal(block_size, use_lock, executable, persistent,
vector, vector_data, heap_region, heap_size,
unit_full);
}
/* Sets the vector data for the lookup vector used by the special heap */
void
special_heap_set_vector_data(void *special, void *vector_data)
{
special_units_t *su = (special_units_t *) special;
special_heap_unit_t *u;
ASSERT(su->heap_areas != NULL);
/* FIXME: more efficient to walk the vector, but no interface
* to set the data: we'd need to expose the iterator index or
* the vmarea struct rather than than the clean copy we have now
*/
for (u = su->top_unit; u != NULL; u = u->next) {
vmvector_modify_data(su->heap_areas, u->alloc_pc,
u->reserved_end_pc, vector_data);
}
}
/* Returns false if the special heap has more than one unit or has a
* non-externally-allocated unit.
* Sets the cur pc for the only unit to end_pc.
*/
bool
special_heap_set_unit_end(void *special, byte *end_pc)
{
special_units_t *su = (special_units_t *) special;
if (su->top_unit->next != NULL ||
SPECIAL_UNIT_HEADER_INLINE(su->top_unit) ||
end_pc < su->top_unit->start_pc ||
end_pc > su->top_unit->end_pc)
return false;
su->top_unit->cur_pc = end_pc;
return true;
}
#ifdef WINDOWS_PC_SAMPLE
static void
special_heap_profile_stop(special_heap_unit_t *u)
{
int sum;
ASSERT(special_heap_profile_enabled());
stop_profile(u->profile);
sum = sum_profile(u->profile);
if (sum > 0) {
mutex_lock(&profile_dump_lock);
print_file(profile_file,
"\nDumping special heap unit profile\n%d hits\n", sum);
dump_profile(profile_file, u->profile);
mutex_unlock(&profile_dump_lock);
}
}
#endif
#if defined(WINDOWS_PC_SAMPLE) && !defined(DEBUG)
/* for fast exit path only, normal path taken care of */
void
special_heap_profile_exit()
{
special_heap_unit_t *u;
special_units_t *su;
ASSERT(special_heap_profile_enabled()); /* will never be compiled in I guess :) */
mutex_lock(&special_units_list_lock);
for (su = special_units_list; su != NULL; su = su->next) {
if (su->use_lock)
mutex_lock(&su->lock);
for (u = su->top_unit; u != NULL; u = u->next) {
if (u->profile != NULL)
special_heap_profile_stop(u);
/* fast exit path: do not bother to free */
}
if (su->use_lock)
mutex_unlock(&su->lock);
}
mutex_unlock(&special_units_list_lock);
}
#endif
void
special_heap_exit(void *special)
{
special_units_t *su = (special_units_t *) special;
special_heap_unit_t *u, *next_u;
#ifdef DEBUG
size_t total_heap_used = 0;
#endif
u = su->top_unit;
while (u != NULL) {
/* Assumption: it's ok to use print_lock even if !su->use_lock */
DOLOG(1, LOG_HEAP|LOG_STATS, {
size_t num_used = u->cur_pc - u->start_pc;
total_heap_used += num_used;
LOG(THREAD_GET, LOG_HEAP|LOG_STATS, 1,
"Heap unit "SZFMT" (size "SZFMT" [/"SZFMT"] KB): used "SZFMT" KB\n",
u->id, (SPECIAL_UNIT_COMMIT_SIZE(u))/1024,
SPECIAL_UNIT_RESERVED_SIZE(u)/1024, num_used/1024);
});
next_u = u->next;
#ifdef WINDOWS_PC_SAMPLE
if (u->profile != NULL) {
ASSERT(special_heap_profile_enabled());
special_heap_profile_stop(u);
free_profile(u->profile);
u->profile = NULL;
}
#endif
STATS_ADD(heap_special_units, -1);
STATS_SUB(heap_special_capacity, SPECIAL_UNIT_COMMIT_SIZE(u));
if (su->heap_areas != NULL) {
vmvector_remove(su->heap_areas, u->alloc_pc, u->reserved_end_pc);
}
if (!SPECIAL_UNIT_HEADER_INLINE(u)) {
HEAP_TYPE_FREE(GLOBAL_DCONTEXT, u, special_heap_unit_t,
ACCT_MEM_MGT, PROTECTED);
/* up to creator to free the heap region */
} else {
release_guarded_real_memory((vm_addr_t)u, SPECIAL_UNIT_RESERVED_SIZE(u),
true/*update DR areas immediately*/, true);
}
u = next_u;
}
#ifdef HEAP_ACCOUNTING
add_heapacct_to_global_stats(&su->acct);
#endif
LOG(THREAD_GET, LOG_HEAP|LOG_STATS, 1, "\tTotal heap used: "SZFMT" KB\n",
total_heap_used/1024);
#if defined(WINDOWS_PC_SAMPLE) && !defined(DEBUG)
if (special_heap_profile_enabled()) {
/* Removed this special_units_t from the master list */
mutex_lock(&special_units_list_lock);
if (special_units_list == su)
special_units_list = su->next;
else {
special_units_t *prev = special_units_list;
ASSERT(prev != NULL);
for (; prev->next != NULL && prev->next != su; prev = prev->next)
;/*nothing*/
ASSERT(prev->next == su);
prev->next = su->next;
}
mutex_unlock(&special_units_list_lock);
}
#endif
if (su->use_lock)
DELETE_LOCK(su->lock);
/* up to caller to free the vector, which is typically multi-heap */
if (su->persistent) {
global_heap_free(su, sizeof(special_units_t) HEAPACCT(ACCT_MEM_MGT));
} else {
nonpersistent_heap_free(GLOBAL_DCONTEXT, su, sizeof(special_units_t)
HEAPACCT(ACCT_MEM_MGT));
}
}
void *
special_heap_calloc(void *special, uint num)
{
#ifdef DEBUG
dcontext_t *dcontext = get_thread_private_dcontext();
#endif
special_units_t *su = (special_units_t *) special;
special_heap_unit_t *u;
void *p = NULL;
bool took_free = false;
ASSERT(num > 0);
if (su->use_lock)
mutex_lock(&su->lock);
u = su->cur_unit;
if (su->free_list != NULL && num == 1) {
p = (void *) su->free_list;
su->free_list = *((heap_pc *)p);
took_free = true;
} else if (su->cfree_list != NULL && num > 1) {
/* FIXME: take a piece of cfree if num == 1?
* seems better to save the bigger pieces
*/
cfree_header_t *cfree = su->cfree_list, *prev = NULL;
while (cfree != NULL && cfree->count < num) {
prev = cfree;
cfree = cfree->next_cfree;
}
if (cfree != NULL) {
ASSERT(cfree->count >= num);
took_free = true;
if (cfree->count == num) {
/* take it out of list altogether */
if (prev == NULL)
su->cfree_list = cfree->next_cfree;
else
prev->next_cfree = cfree->next_cfree;
p = (void *) cfree;
} else if (cfree->count == num+1) {
/* add single-size piece to normal free list */
heap_pc tail = ((heap_pc) cfree) + num * su->block_size;
*((heap_pc *)tail) = su->free_list;
su->free_list = tail;
p = (void *) cfree;
} else {
/* if take tail don't have to change free list ptrs at all */
cfree->count -= num;
p = (void *) (((heap_pc) cfree) + (cfree->count - num) * su->block_size);
}
}
}
if (!took_free) {
/* no free blocks, grab a new one */
if (u->cur_pc + su->block_size*num > u->end_pc ||
POINTER_OVERFLOW_ON_ADD(u->cur_pc, su->block_size*num)) {
/* simply extend commitment, if possible */
DEBUG_DECLARE(size_t pre_commit_size = SPECIAL_UNIT_COMMIT_SIZE(u);)
special_unit_extend_commitment(u, su->block_size*num, get_prot(su));
STATS_ADD_PEAK(heap_special_capacity,
SPECIAL_UNIT_COMMIT_SIZE(u) - pre_commit_size);
/* check again after extending commit */
if (u->cur_pc + su->block_size*num > u->end_pc ||
POINTER_OVERFLOW_ON_ADD(u->cur_pc, su->block_size*num)) {
/* no room, need new unit */
special_heap_unit_t *new_unit;
special_heap_unit_t *prev = su->top_unit;
size_t size = UNITALLOC(u);
while (prev->next != NULL)
prev = prev->next;
/* create new unit double size of old unit (until hit max size) */
if (size*2 <= HEAP_UNIT_MAX_SIZE)
size *= 2;
/* we don't support arbitrarily long sequences */
ASSERT(su->block_size * num < size);
new_unit = special_heap_create_unit(su, NULL, size, false/*empty*/);
prev->next = new_unit;
if (su->use_lock) {
/* if synch bad so is printing */
LOG(THREAD, LOG_HEAP, 3, "\tCreating new heap unit %d\n",
new_unit->id);
}
su->cur_unit = new_unit;
u = new_unit;
ASSERT(u->cur_pc + su->block_size*num <= u->end_pc &&
!POINTER_OVERFLOW_ON_ADD(u->cur_pc, su->block_size*num));
}
}
p = (void *) u->cur_pc;
u->cur_pc += su->block_size*num;
ACCOUNT_FOR_ALLOC(alloc_new, su, ACCT_SPECIAL,
su->block_size*num, su->block_size*num);
} else {
ACCOUNT_FOR_ALLOC(alloc_reuse, su, ACCT_SPECIAL,
su->block_size*num, su->block_size*num);
}
if (su->use_lock)
mutex_unlock(&su->lock);
#ifdef DEBUG_MEMORY
DOCHECK(CHKLVL_MEMFILL, memset(p, HEAP_ALLOCATED_BYTE, su->block_size*num););
#endif
ASSERT(p != NULL);
return (void*)p;
}
void *
special_heap_alloc(void *special)
{
return special_heap_calloc(special, 1);
}
void
special_heap_cfree(void *special, void *p, uint num)
{
special_units_t *su = (special_units_t *) special;
ASSERT(num > 0);
ASSERT(p != NULL);
/* Allow freeing while iterating w/o deadlock (iterator holds lock) */
ASSERT(!su->in_iterator || OWN_MUTEX(&su->lock));
if (su->use_lock && !su->in_iterator)
mutex_lock(&su->lock);
#ifdef DEBUG_MEMORY
/* FIXME: ensure that p is in allocated state */
DOCHECK(CHKLVL_MEMFILL, memset(p, HEAP_UNALLOCATED_BYTE, su->block_size*num););
#endif
if (num == 1) {
/* write next pointer */
*((heap_pc *)p) = su->free_list;
su->free_list = (heap_pc)p;
} else {
cfree_header_t *cfree = (cfree_header_t *) p;
cfree->next_cfree = su->cfree_list;
cfree->count = num;
su->cfree_list = cfree;
}
#ifdef HEAP_ACCOUNTING
ACCOUNT_FOR_FREE(su, ACCT_SPECIAL, su->block_size*num);
#endif
if (su->use_lock && !su->in_iterator)
mutex_unlock(&su->lock);
}
void
special_heap_free(void *special, void *p)
{
special_heap_cfree(special, p, 1);
}
/* Special heap iterator. Initialized with special_heap_iterator_start(), which
* grabs the heap lock (regardless of whether synch is used for allocs), and
* destroyed with special_heap_iterator_stop() to release the lock.
* If the special heap uses no lock for alloc, it is up to the caller
* to prevent race conditions causing problems.
* Accessor special_heap_iterator_next() should be called only when
* predicate special_heap_iterator_hasnext() is true.
* Any mutation of the heap while iterating will result in a deadlock
* for heaps that use locks for alloc, except for individual freeing,
* which will proceed w/o trying to grab the lock a second time.
* FIXME: could generalize to regular heaps if a use arises.
*/
void
special_heap_iterator_start(void *heap, special_heap_iterator_t *shi)
{
special_units_t *su = (special_units_t *) heap;
ASSERT(heap != NULL);
ASSERT(shi != NULL);
mutex_lock(&su->lock);
shi->heap = heap;
shi->next_unit = (void *) su->top_unit;
su->in_iterator = true;
}
bool
special_heap_iterator_hasnext(special_heap_iterator_t *shi)
{
ASSERT(shi != NULL);
DOCHECK(1, {
special_units_t *su = (special_units_t *) shi->heap;
ASSERT(su != NULL);
ASSERT_OWN_MUTEX(true, &su->lock);
});
return (shi->next_unit != NULL);
}
/* Iterator accessor:
* Has to be initialized with special_heap_iterator_start, and should be
* called only when special_heap_iterator_hasnext() is true.
* Sets the area boundaries in area_start and area_end.
*/
void
special_heap_iterator_next(special_heap_iterator_t *shi /* IN/OUT */,
app_pc *heap_start /* OUT */, app_pc *heap_end /* OUT */)
{
special_units_t *su;
special_heap_unit_t *u;
ASSERT(shi != NULL);
su = (special_units_t *) shi->heap;
ASSERT(su != NULL);
ASSERT_OWN_MUTEX(true, &su->lock);
u = (special_heap_unit_t *) shi->next_unit;
ASSERT(u != NULL);
if (u != NULL) { /* caller error, but paranoid */
if (heap_start != NULL)
*heap_start = u->start_pc;
ASSERT(u->cur_pc <= u->end_pc);
if (heap_end != NULL)
*heap_end = u->cur_pc;
shi->next_unit = (void *) u->next;
}
}
void
special_heap_iterator_stop(special_heap_iterator_t *shi)
{
special_units_t *su;
ASSERT(shi != NULL);
su = (special_units_t *) shi->heap;
ASSERT(su != NULL);
ASSERT_OWN_MUTEX(true, &su->lock);
su->in_iterator = false;
mutex_unlock(&su->lock);
DODEBUG({
shi->heap = NULL;
shi->next_unit = NULL;
});
}
#if defined(DEBUG) && defined(HOT_PATCHING_INTERFACE)
/* We leak hotp trampolines as part of fix for case 9593; so, during a detach
* we can't delete the trampoline heap. However if that heap's lock isn't
* deleted, we'll assert. This routine is used only for that. Normally, we
* should call special_heap_exit() which deletes the lock. */
void
special_heap_delete_lock(void *special)
{
special_units_t *su = (special_units_t *)special;
/* No one calls this routine unless they have a lock to delete. */
ASSERT(su != NULL);
if (su == NULL)
return;
ASSERT(su->use_lock);
if (su->use_lock)
DELETE_LOCK(su->lock);
}
#endif
/*----------------------------------------------------------------------------*/
#ifdef WINDOWS /* currently not used on linux */
/* Landing pads (introduced as part of work for PR 250294). */
/* landing_pad_areas is a vmvector made up of regions of memory called
* landing pad areas, each of which contains multiple landing pads. Landing
* pads are small trampolines used to jump from the hook point to the main
* trampoline. This is used in both 32-bit and 64-bit DR. In both cases it
* will handle the problem of hook chaining by 3rd party software and us having
* to release our hooks (we'll nop the landing pad and free the trampoline).
* In 64-bit it also solves the problem of reachability of the 5-byte rel jmp
* we use for hooking, i.e., that 5-byte rel jmp may not reach the main
* trampoline in DR heap. We have to maintain the hook a 5-byte jmp because
* hotp_only assumes it (see PR 250294).
*
* A landing pad will have nothing more than a jump (5-byte rel for 32-bit DR
* and 64-bit abs ind jmp for 64-bit DR) to the trampoline and a 5-byte rel jmp
* back to the next instruction after the hook, plus the displaced app instrs.
*
* To handle hook chaining landing pads won't be released till process exit
* (not on a detach), their first jump will just be nop'ed. As landing pads
* aren't released till exit, all landing pads are just incrementally allocated
* in a landing pad area.
*
* Note: Landing pad areas don't necessarily have to fall within the vm_reserve
* region or capacity, so aren't accounted by our vmm.
*
* Note: If in future other needs for such region specific allocation should
* arise, then we should convert this into special_heap_alloc_in_region(). For
* now, landing pads are the only consumers, so was decided to be acceptable.
*
* See win32/callback.c for emit_landing_pad_code() and landing pad usage.
*/
typedef struct {
byte *start; /* start of reserved region */
byte *end; /* end of reserved region */
byte *commit_end; /* end of committed memory in the reserved region */
byte *cur_ptr; /* pointer to next allocatable landing pad memory */
bool allocated; /* allocated, or stolen from an app dll? */
} landing_pad_area_t;
/* Allocates a landing pad so that a hook inserted at addr_to_hook can reach
* its trampoline via the landing pad. The landing pad will reachable by a
* 32-bit relative jmp from addr_to_hook.
* Note: we may want to generalize this at some point such that the size of the
* landing pad is passed as an argument.
*
* For Windows we assume that landing_pads_to_executable_areas(true) will be
* called once landing pads are finished being created.
*/
byte *
alloc_landing_pad(app_pc addr_to_hook)
{
app_pc hook_region_start, hook_region_end;
app_pc alloc_region_start, alloc_region_end;
app_pc lpad_area_start = NULL, lpad_area_end;
app_pc lpad = NULL;
landing_pad_area_t *lpad_area = NULL;
/* Allocate the landing pad area such that any hook from within the module
* or memory region containing addr_to_hook can use the same area for a
* landing pad. Makes it more efficient. */
hook_region_start = get_allocation_base(addr_to_hook);
if (hook_region_start == NULL) { /* to support raw virtual address hooks */
ASSERT_CURIOSITY("trying to hook raw or unallocated memory?");
hook_region_start = addr_to_hook;
hook_region_end = addr_to_hook;
} else {
hook_region_end = hook_region_start +
get_allocation_size(hook_region_start, NULL);
ASSERT(hook_region_end > hook_region_start); /* check overflow */
/* If region size is > 2 GB, then it isn't an image; PE32{,+} restrict
* images to 2 GB. Also, if region is > 2 GB the reachability macros
* called below will return a region smaller (and with start and end
* inverted) than the region from which the reachability is desired,
* i.e., some of the areas in [hook_region_start, hook_region_end)
* won't be able to reach the region computed.
*
* A better choice is to pick something smaller (100 MB) because if the
* region is close to 2 GB in size then we might not be able to
* allocate memory for a landing pad that is reachable.
*/
if (hook_region_end - hook_region_start > 100 * 1024 * 1024) {
/* Try a smaller region of 100 MB around the address to hook. */
ASSERT_CURIOSITY(false && "seeing patch region > 100 MB - DGC?");
hook_region_start = MIN(addr_to_hook, MAX(hook_region_start,
addr_to_hook - 50 * 1024 * 1024));
hook_region_end = MAX(addr_to_hook, MIN(hook_region_end,
addr_to_hook + 50 * 1024 * 1024));
}
}
/* Define the region that can be reached from anywhere within the
* hook region with a 32-bit rel jmp.
*/
alloc_region_start = REACHABLE_32BIT_START(hook_region_start,
hook_region_end);
alloc_region_end = REACHABLE_32BIT_END(hook_region_start, hook_region_end);
ASSERT(alloc_region_start < alloc_region_end);
/* Check if there is an existing landing pad area within the reachable
* region for the hook location. If so use it, else allocate one.
*/
write_lock(&landing_pad_areas->lock);
if (vmvector_overlap(landing_pad_areas, alloc_region_start,
alloc_region_end)) {
/* Now we have to get that landing pad area that is FULLY contained
* within alloc_region_start and alloc_region_end. If a landing pad
* area is only partially within the alloc region, then a landing pad
* created there won't be able to reach the addr_to_hook. BTW, that
* landing pad area should have enough space to allocate a landing pad!
* If these conditions are met allocate a landing pad.
*/
vmvector_iterator_t lpad_area_iter;
vmvector_iterator_start(landing_pad_areas, &lpad_area_iter);
while (vmvector_iterator_hasnext(&lpad_area_iter)) {
lpad_area = vmvector_iterator_next(&lpad_area_iter,
&lpad_area_start,
&lpad_area_end);
if (lpad_area_start < alloc_region_end &&
lpad_area_end > alloc_region_start &&
(lpad_area->cur_ptr + LANDING_PAD_SIZE) < lpad_area_end) {
/* See if enough memory in this landing pad area has been
* committed, if not commit more memory.
*/
if ((lpad_area->cur_ptr + LANDING_PAD_SIZE) >=
lpad_area->commit_end) {
ASSERT(lpad_area->allocated);
extend_commitment(lpad_area->commit_end, PAGE_SIZE,
MEMPROT_READ|MEMPROT_EXEC,
false /* not initial commit */);
lpad_area->commit_end += PAGE_SIZE;
}
/* Update the current pointer for the landing pad area, i.e.,
* allocate the landing pad.
*/
lpad = lpad_area->cur_ptr;
lpad_area->cur_ptr += LANDING_PAD_SIZE;
break;
}
}
vmvector_iterator_stop(&lpad_area_iter);
}
/* If a landing pad area wasn't found because there wasn't any in the
* allocation region or none fully contained within the allocation region,
* then create a new one within the allocation region. Then allocate a
* landing pad in it.
*/
if (lpad == NULL) {
bool allocated = true;
heap_error_code_t heap_error;
lpad_area_end = NULL;
lpad_area_start = os_heap_reserve_in_region
((void *)ALIGN_FORWARD(alloc_region_start, PAGE_SIZE),
(void *)ALIGN_BACKWARD(alloc_region_end, PAGE_SIZE),
LANDING_PAD_AREA_SIZE, &heap_error, true/*+x*/);
if (lpad_area_start == NULL ||
heap_error == HEAP_ERROR_CANT_RESERVE_IN_REGION) {
/* Should retry with using just the aligned target address - we may
* have made the region so large that there's nothing nearby to
* reserve.
*/
lpad_area_start = os_heap_reserve(
(void *)ALIGN_FORWARD(addr_to_hook,
LANDING_PAD_AREA_SIZE),
LANDING_PAD_AREA_SIZE, &heap_error, true/*+x*/);
#ifdef WINDOWS
if (lpad_area_start == NULL &&
/* We can only do this once w/ current interface.
* XXX: support multiple "allocs" inside libs.
*/
vmvector_empty(landing_pad_areas) &&
os_find_free_code_space_in_libs(&lpad_area_start, &lpad_area_end)) {
if (lpad_area_end - lpad_area_start >= LANDING_PAD_SIZE &&
/* Mark writable until we're done creating landing pads */
make_hookable(lpad_area_start, lpad_area_end - lpad_area_start,
NULL)) {
/* Let's take it */
allocated = false;
/* We assume that landing_pads_to_executable_areas(true) will be
* called once landing pads are finished being created and we
* can restore to +rx there.
*/
lpad_temp_writable_start = lpad_area_start;
lpad_temp_writable_size = lpad_area_end - lpad_area_start;
} else
lpad_area_start = NULL; /* not big enough */
}
#endif
if (lpad_area_start == NULL) {
/* Even at startup when there will be enough memory,
* theoretically 2 GB of dlls might get packed together before
* we get control (very unlikely), so we can fail. If it does,
* then say 'oom' and exit.
*/
SYSLOG_INTERNAL_WARNING("unable to reserve memory for landing pads");
report_low_on_memory(OOM_RESERVE, heap_error);
}
}
/* Allocate the landing pad area as rx, allocate a landing pad in it
* and add it to landing_pad_areas vector. Note, we only commit 4k
* initially even though we reserve 64k (LANDING_PAD_AREA_SIZE), to
* avoid wastage.
*/
if (allocated) {
extend_commitment(lpad_area_start, PAGE_SIZE, MEMPROT_READ|MEMPROT_EXEC,
true /* initial commit */);
}
lpad_area = HEAP_TYPE_ALLOC(GLOBAL_DCONTEXT, landing_pad_area_t,
ACCT_VMAREAS, PROTECTED);
lpad_area->start = lpad_area_start;
lpad_area->end = (lpad_area_end == NULL ?
lpad_area_start + LANDING_PAD_AREA_SIZE : lpad_area_end);
lpad_area->commit_end = lpad_area_start + PAGE_SIZE;
lpad_area->cur_ptr = lpad_area_start;
lpad_area->allocated = allocated;
lpad = lpad_area->cur_ptr;
lpad_area->cur_ptr += LANDING_PAD_SIZE;
vmvector_add(landing_pad_areas, lpad_area->start,
lpad_area->end, lpad_area);
STATS_INC(num_landing_pad_areas);
}
/* Landing pads aren't added to executable_areas here because not all
* landing pads should be added. Only the ones used for DR hooks should be
* added to executable areas (which is done using
* landing_pads_to_executable_areas() at the end of inserting DR hooks).
* hotp_only related landing pads shouldn't be added to executable areas as
* their trampolines aren't added to executable_areas. This is why landing
* pads aren't added to executable_areas here, at the point of allocation.
*/
LOG(GLOBAL, LOG_ALL, 3, "%s: used "PIFX" bytes in "PFX"-"PFX"\n", __FUNCTION__,
lpad_area->cur_ptr - lpad_area->start, lpad_area->start, lpad_area->end);
/* Boundary check to make sure the allocation is within the landing pad area. */
ASSERT(lpad_area->cur_ptr <= lpad_area->end);
write_unlock(&landing_pad_areas->lock);
return lpad;
}
/* Attempts to save space in the landing pad region by trimming the most
* recently allocated landing pad to the actual space used.
* Will fail if another landing pad was allocated between lpad_start
* being allocated and this routine being called.
*/
bool
trim_landing_pad(byte *lpad_start, size_t space_used)
{
landing_pad_area_t *lpad_area = NULL;
bool res = false;
write_lock(&landing_pad_areas->lock);
if (vmvector_lookup_data(landing_pad_areas, lpad_start, NULL,
NULL, &lpad_area)) {
if (lpad_start == lpad_area->cur_ptr - LANDING_PAD_SIZE) {
lpad_area->cur_ptr -= (LANDING_PAD_SIZE - space_used);
res = true;
}
}
write_unlock(&landing_pad_areas->lock);
return res;
}
/* Adds or removes all landing pads from executable_areas by adding whole
* landing pad areas. This is done to prevent bb building from considering
* landing pads to be selfmod code; as such, these don't have to be
* {add,remov}ed from executable_areas for hotp_only or for thin_client mode.
*/
void
landing_pads_to_executable_areas(bool add)
{
vmvector_iterator_t lpad_area_iter;
app_pc lpad_area_start, lpad_area_end;
DEBUG_DECLARE(landing_pad_area_t *lpad_area;)
uint lpad_area_size;
if (RUNNING_WITHOUT_CODE_CACHE())
return;
#ifdef WINDOWS
if (add && lpad_temp_writable_start != NULL) {
make_unhookable(lpad_temp_writable_start, lpad_temp_writable_size, true);
lpad_temp_writable_start = NULL;
}
#endif
/* With code cache, there should be only one landing pad area, just for
* dr hooks in ntdll. For 64-bit, the image entry hook will result in a
* new landing pad.
*/
IF_X64_ELSE(,ASSERT(landing_pad_areas->length == 1);)
/* Just to be safe, walk through all areas in release build. */
vmvector_iterator_start(landing_pad_areas, &lpad_area_iter);
while (vmvector_iterator_hasnext(&lpad_area_iter)) {
DEBUG_DECLARE(lpad_area = )
vmvector_iterator_next(&lpad_area_iter, &lpad_area_start,
&lpad_area_end);
lpad_area_size = (uint)(lpad_area_end - lpad_area_start);
ASSERT(lpad_area_size <= LANDING_PAD_AREA_SIZE);
/* Current ptr should be within area. */
ASSERT(lpad_area->cur_ptr < lpad_area_end);
if (add) {
add_executable_region(lpad_area_start, lpad_area_size
_IF_DEBUG("add landing pad areas after inserting dr hooks"));
} else {
remove_executable_region(lpad_area_start, lpad_area_size,
false /* no lock */);
}
}
vmvector_iterator_stop(&lpad_area_iter);
}
/* Delete landing_pad_areas and the landing_pad_area_t allocated for each
* landing pad area. However, release all landing pads only on process exit;
* for detach leave the landing pads in (in case some one hooks after us they
* shouldn't crash if they chain correctly).
*/
static void
release_landing_pad_mem(void)
{
vmvector_iterator_t lpad_area_iter;
app_pc lpad_area_start, lpad_area_end;
landing_pad_area_t *lpad_area;
heap_error_code_t heap_error;
vmvector_iterator_start(landing_pad_areas, &lpad_area_iter);
while (vmvector_iterator_hasnext(&lpad_area_iter)) {
bool allocated;
lpad_area = vmvector_iterator_next(&lpad_area_iter, &lpad_area_start,
&lpad_area_end);
allocated = lpad_area->allocated;
HEAP_TYPE_FREE(GLOBAL_DCONTEXT, lpad_area, landing_pad_area_t,
ACCT_VMAREAS, PROTECTED);
if (!doing_detach && /* On normal exit release the landing pads. */
allocated)
os_heap_free(lpad_area_start, LANDING_PAD_AREA_SIZE, &heap_error);
}
vmvector_iterator_stop(&lpad_area_iter);
vmvector_delete_vector(GLOBAL_DCONTEXT, landing_pad_areas);
}
#endif /* WINDOWS */
/*----------------------------------------------------------------------------*/