4e8ee9ab65
ratio parameter
381 lines
16 KiB
C
381 lines
16 KiB
C
/*
|
|
* This file is part of the COMROGUE Operating System for Raspberry Pi
|
|
*
|
|
* Copyright (c) 2013, Eric J. Bowersox / Erbosoft Enterprises
|
|
* All rights reserved.
|
|
*
|
|
* This program is free for commercial and non-commercial use as long as the following conditions are
|
|
* adhered to.
|
|
*
|
|
* Copyright in this file remains Eric J. Bowersox and/or Erbosoft, and as such any copyright notices
|
|
* in the code are not to be removed.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without modification, are permitted
|
|
* provided that the following conditions are met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright notice, this list of conditions and
|
|
* the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and
|
|
* the following disclaimer in the documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED
|
|
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A
|
|
* PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR
|
|
* ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
|
|
* TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
|
* NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* "Raspberry Pi" is a trademark of the Raspberry Pi Foundation.
|
|
*/
|
|
/*
|
|
* This code is based on/inspired by jemalloc-3.3.1. Please see LICENSE.jemalloc for further details.
|
|
*/
|
|
#ifndef __HEAP_INTERNALS_H_INCLUDED
|
|
#define __HEAP_INTERNALS_H_INCLUDED
|
|
|
|
#ifdef __COMROGUE_INTERNALS__
|
|
|
|
#ifndef __ASM__
|
|
|
|
#include <stdarg.h>
|
|
#include <comrogue/compiler_macros.h>
|
|
#include <comrogue/types.h>
|
|
#include <comrogue/objectbase.h>
|
|
#include <comrogue/connpoint.h>
|
|
#include <comrogue/allocator.h>
|
|
#include <comrogue/stream.h>
|
|
#include <comrogue/mutex.h>
|
|
#include <comrogue/heap.h>
|
|
#include <comrogue/objhelp.h>
|
|
#include <comrogue/internals/dlist.h>
|
|
#include <comrogue/internals/rbtree.h>
|
|
#include <comrogue/internals/seg.h>
|
|
|
|
/*--------------------------------------------------------------
|
|
* Radix tree implementation for keeping track of memory chunks
|
|
*--------------------------------------------------------------
|
|
*/
|
|
|
|
#define RTREE_NODESIZE (1U << 14) /* node size for tree */
|
|
|
|
typedef struct tagMEMRTREE
|
|
{
|
|
IMutex *pmtx; /* mutex for tree */
|
|
PPVOID ppRoot; /* tree root */
|
|
UINT32 uiHeight; /* tree height */
|
|
UINT32 auiLevel2Bits[0]; /* bits at level 2 - dynamically sized */
|
|
} MEMRTREE, *PMEMRTREE;
|
|
|
|
/*-------------------
|
|
* Extent management
|
|
*-------------------
|
|
*/
|
|
|
|
/* Tree of extents managed by the heap management code. */
|
|
typedef struct tagEXTENT_NODE
|
|
{
|
|
RBTREENODE rbtnSizeAddress; /* tree node for size and address ordering */
|
|
RBTREENODE rbtnAddress; /* tree node for address ordering */
|
|
/* TODO prof_ctx? */
|
|
PVOID pv; /* base pointer to region */
|
|
SIZE_T sz; /* size of region */
|
|
BOOL fZeroed; /* is this extent zeroed? */
|
|
} EXTENT_NODE, *PEXTENT_NODE;
|
|
typedef PEXTENT_NODE *PPEXTENT_NODE;
|
|
|
|
/*---------------------
|
|
* Bitmap declarations
|
|
*---------------------
|
|
*/
|
|
|
|
/* Maximum number of regions per run */
|
|
#define LG_RUN_MAXREGS 11
|
|
#define RUN_MAXREGS (1U << LG_RUN_MAXREGS)
|
|
|
|
/* Maximum bitmap count */
|
|
#define LG_BITMAP_MAXBITS LG_RUN_MAXREGS
|
|
|
|
typedef UINT32 BITMAP; /* bitmap type definition */
|
|
#define LG_SIZEOF_BITMAP LOG_UINTSIZE
|
|
|
|
/* Number of bits per group */
|
|
#define LG_BITMAP_GROUP_NBITS (LG_SIZEOF_BITMAP + 3)
|
|
#define BITMAP_GROUP_NBITS (1U << LG_BITMAP_GROUP_NBITS)
|
|
#define BITMAP_GROUP_NBITS_MASK (BITMAP_GROUP_NBITS - 1)
|
|
|
|
/* Maximum number of bitmap levels */
|
|
#define BITMAP_MAX_LEVELS \
|
|
((LG_BITMAP_MAXBITS / LG_SIZEOF_BITMAP) + !!(LG_BITMAP_MAXBITS % LG_SIZEOF_BITMAP))
|
|
|
|
/* Bitmap level structure */
|
|
typedef struct tagBITMAPLEVEL
|
|
{
|
|
SIZE_T ofsGroup; /* offset of groups for this level within array */
|
|
} BITMAPLEVEL, *PBITMAPLEVEL;
|
|
|
|
/* Bitmap information structure */
|
|
typedef struct tagBITMAPINFO
|
|
{
|
|
SIZE_T cBits; /* number of bits in bitmap */
|
|
UINT32 nLevels; /* number of levels required for bits */
|
|
BITMAPLEVEL aLevels[BITMAP_MAX_LEVELS + 1]; /* the levels - only first (nLevels + 1) used */
|
|
} BITMAPINFO, *PBITMAPINFO;
|
|
|
|
/*---------------------------------
|
|
* Thread-level cache declarations
|
|
*---------------------------------
|
|
*/
|
|
|
|
/* statistics per cache bin */
|
|
typedef struct tagTCACHEBINSTATS
|
|
{
|
|
UINT64 nRequests; /* number of requests in this particular bin */
|
|
} TCACHEBINSTATS, *PTCACHEBINSTATS;
|
|
|
|
/* single bin of the cache */
|
|
typedef struct tagTCACHEBIN
|
|
{
|
|
TCACHEBINSTATS stats; /* statistics for this bin */
|
|
INT32 nLowWatermark; /* minimum number cached since last GC */
|
|
UINT32 cbitFill; /* fill level */
|
|
UINT32 nCached; /* number of cached objects */
|
|
PPVOID ppvAvail; /* stack of cached objects */
|
|
} TCACHEBIN, *PTCACHEBIN;
|
|
|
|
typedef struct tagARENA ARENA, *PARENA; /* forward declaration */
|
|
|
|
/* thread cache */
|
|
typedef struct tagTCACHE
|
|
{
|
|
DLIST_FIELD_DECLARE(struct tagTCACHE, link); /* link aggregator */
|
|
UINT64 cbProfAccum; /* accumulated bytes */
|
|
PARENA parena; /* this thread's arena */
|
|
UINT32 cEvents; /* event count since incremental GC */
|
|
UINT32 ndxNextGCBin; /* next bin to be GC'd */
|
|
TCACHEBIN aBins[0]; /* cache bins (dynamically sized) */
|
|
} TCACHE, *PTCACHE;
|
|
|
|
/*------------------------
|
|
* Arena data definitions
|
|
*------------------------
|
|
*/
|
|
|
|
/* Chunk map, each element corresponds to one page within the chunk */
|
|
typedef struct tagARENACHUNKMAP
|
|
{
|
|
union
|
|
{
|
|
RBTREENODE rbtn; /* tree of runs */
|
|
DLIST_FIELD_DECLARE(struct tagARENACHUNKMAP, link); /* list of runs in purgatory */
|
|
} u;
|
|
SIZE_T bits; /* run address and various flags */
|
|
} ARENACHUNKMAP, *PARENACHUNKMAP;
|
|
|
|
#define CHUNK_MAP_BININD_SHIFT 4 /* shift count for bin index mask */
|
|
#define BININD_INVALID ((SIZE_T)0xFFU) /* invalid bin index */
|
|
#define CHUNK_MAP_BININD_MASK ((SIZE_T)0xFF0U) /* bin index mask */
|
|
#define CHUNK_MAP_BININD_INVALID CHUNK_MAP_BININD_MASK /* invalid bin marker */
|
|
#define CHUNK_MAP_FLAGS_MASK ((SIZE_T)0xCU) /* flag bits mask */
|
|
#define CHUNK_MAP_DIRTY ((SIZE_T)0x8U) /* dirty flag */
|
|
#define CHUNK_MAP_UNZEROED ((SIZE_T)0x4U) /* non-zeroed flag */
|
|
#define CHUNK_MAP_LARGE ((SIZE_T)0x2U) /* large allocation flag */
|
|
#define CHUNK_MAP_ALLOCATED ((SIZE_T)0x1U) /* allocated flag */
|
|
#define CHUNK_MAP_KEY CHUNK_MAP_ALLOCATED
|
|
|
|
/* chunk header within an arena */
|
|
typedef struct tagARENACHUNK
|
|
{
|
|
PARENA parena; /* arena that owns the chunk */
|
|
RBTREENODE rbtnDirty; /* linkage for tree of chunks with dirty runs */
|
|
SIZE_T cpgDirty; /* number of dirty pages */
|
|
SIZE_T cAvailRuns; /* number of available runs */
|
|
SIZE_T cAvailRunAdjacent; /* number of available run adjacencies */
|
|
ARENACHUNKMAP aMaps[0]; /* map of pages within chunk */
|
|
} ARENACHUNK, *PARENACHUNK;
|
|
|
|
/* large allocation statistics */
|
|
typedef struct tagMALLOCLARGESTATS
|
|
{
|
|
UINT64 nMalloc; /* number of allocation requests */
|
|
UINT64 nDalloc; /* number of deallocation requests */
|
|
UINT64 nRequests; /* number of allocation requests */
|
|
SIZE_T cRuns; /* count of runs of this size class */
|
|
} MALLOCLARGESTATS, *PMALLOCLARGESTATS;
|
|
|
|
/* Arena statistics data. */
|
|
typedef struct tagARENASTATS
|
|
{
|
|
SIZE_T cbMapped; /* number of bytes currently mapped */
|
|
UINT64 cPurges; /* number of purge sweeps made */
|
|
UINT64 cAdvise; /* number of memory advise calls made */
|
|
UINT64 cPagesPurged; /* number of pages purged */
|
|
SIZE_T cbAllocatedLarge; /* number of bytes of large allocations */
|
|
UINT64 cLargeMalloc; /* number of large allocations */
|
|
UINT64 cLargeDalloc; /* number of large deallocations */
|
|
UINT64 cLargeRequests; /* number of large allocation requests */
|
|
PMALLOCLARGESTATS amls; /* array of stat elements, one per size class */
|
|
} ARENASTATS, *PARENASTATS;
|
|
|
|
struct tagARENABININFO
|
|
{
|
|
SIZE_T cbRegions; /* size of regions in a run */
|
|
SIZE_T cbRedzone; /* size of the red zone */
|
|
SIZE_T cbInterval; /* interval between regions */
|
|
SIZE_T cbRunSize; /* total size of a run for this size class */
|
|
UINT32 nRegions; /* number of regions in a run for this size class */
|
|
UINT32 ofsBitmap; /* offset of bitmap element in run header */
|
|
BITMAPINFO bitmapinfo; /* manipulates bitmaps associated with this bin's runs */
|
|
UINT32 ofsCtx0; /* offset of context in run header, or 0 */
|
|
UINT32 ofsRegion0; /* offse of first region in a run for size class */
|
|
} ARENABININFO, *PARENABININFO;
|
|
|
|
/* The actual arena definition. */
|
|
struct tagARENA
|
|
{
|
|
UINT32 nIndex; /* index of this arena within array */
|
|
UINT32 nThreads; /* number of threads assigned to this arena */
|
|
IMutex *pmtxLock; /* arena lock */
|
|
ARENASTATS stats; /* arena statistics */
|
|
DLIST_HEAD_DECLARE(TCACHE, dlistTCache); /* list of tcaches for threads in arena */
|
|
UINT64 cbProfAccum; /* accumulated bytes */
|
|
RBTREE rbtDirtyChunks; /* tree of dirty page-containing chunks */
|
|
};
|
|
|
|
/*----------------------------------
|
|
* The actual heap data declaration
|
|
*----------------------------------
|
|
*/
|
|
|
|
typedef struct tagHEAPDATA {
|
|
IMalloc mallocInterface; /* pointer to IMalloc interface - MUST BE FIRST! */
|
|
IConnectionPointContainer cpContainerInterface; /* pointer to IConnectionPointContainer interface */
|
|
IHeapConfiguration heapConfInterface; /* pointer to IHeapConfiguration interface */
|
|
UINT32 uiRefCount; /* reference count */
|
|
UINT32 uiFlags; /* flags word */
|
|
PFNRAWHEAPDATAFREE pfnFreeRawHeapData; /* pointer to function that frees the raw heap data, if any */
|
|
PFNHEAPABORT pfnAbort; /* pointer to abort function */
|
|
PVOID pvAbortArg; /* argument to abort function */
|
|
IChunkAllocator *pChunkAllocator; /* chunk allocator pointer */
|
|
IMutexFactory *pMutexFactory; /* mutex factory pointer */
|
|
FIXEDCPDATA fcpMallocSpy; /* connection point for IMallocSpy */
|
|
FIXEDCPDATA fcpSequentialStream; /* connection point for ISequentialStream for debugging */
|
|
IMallocSpy *pMallocSpy; /* IMallocSpy interface for the allocator */
|
|
ISequentialStream *pDebugStream; /* debugging output stream */
|
|
UINT32 nChunkBits; /* number of bits in a chunk */
|
|
UINT32 szChunk; /* size of a chunk */
|
|
UINT32 uiChunkSizeMask; /* bitmask for a chunk */
|
|
UINT32 cpgChunk; /* number of pages in a chunk */
|
|
SSIZE_T cbActiveDirtyRatio; /* active/dirty ratio parameter */
|
|
IMutex *pmtxChunks; /* chunks mutex */
|
|
RBTREE rbtExtSizeAddr; /* tree ordering extents by size and address */
|
|
RBTREE rbtExtAddr; /* tree ordering extents by address */
|
|
PMEMRTREE prtChunks; /* radix tree containing all chunk values */
|
|
IMutex *pmtxBase; /* base mutex */
|
|
PVOID pvBasePages; /* pages being used for internal memory allocation */
|
|
PVOID pvBaseNext; /* next allocation location */
|
|
PVOID pvBasePast; /* address immediately past pvBasePages */
|
|
PEXTENT_NODE pexnBaseNodes; /* pointer to base nodes */
|
|
} HEAPDATA, *PHEAPDATA;
|
|
|
|
/*---------------------------------
|
|
* Utility and debugging functions
|
|
*---------------------------------
|
|
*/
|
|
|
|
/* Get nearest aligned address at or below a. */
|
|
#define ALIGNMENT_ADDR2BASE(a, alignment) ((PVOID)(((UINT_PTR)(a)) & ~(alignment)))
|
|
|
|
/* Get offset between a and ALIGNMENT_ADDR2BASE(a, alignment). */
|
|
#define ALIGNMENT_ADDR2OFFSET(a, alignment) ((SIZE_T)(((UINT_PTR)(a)) & (alignment)))
|
|
|
|
/* Returns the smallest alignment multiple greater than sz. */
|
|
#define ALIGNMENT_CEILING(sz, alignment) (((sz) + (alignment)) & ~(alignment))
|
|
|
|
CDECL_BEGIN
|
|
|
|
extern void _HeapDbgWrite(PHEAPDATA phd, PCSTR sz);
|
|
extern void _HeapPrintf(PHEAPDATA phd, PCSTR szFormat, ...);
|
|
extern void _HeapAssertFailed(PHEAPDATA phd, PCSTR szFile, INT32 nLine);
|
|
extern SIZE_T _HeapPow2Ceiling(SIZE_T x);
|
|
|
|
CDECL_END
|
|
|
|
#define _H_THIS_FILE __FILE__
|
|
#define _DECLARE_H_THIS_FILE static const char SEG_RODATA _H_THIS_FILE[] = __FILE__;
|
|
|
|
#define _H_ASSERT(phd, expr) ((expr) ? (void)0 : _HeapAssertFailed(phd, _H_THIS_FILE, __LINE__))
|
|
|
|
/*---------------------------------
|
|
* Radix tree management functions
|
|
*---------------------------------
|
|
*/
|
|
|
|
CDECL_BEGIN
|
|
|
|
extern PMEMRTREE _HeapRTreeNew(PHEAPDATA phd, UINT32 cBits);
|
|
extern void _HeapRTreeDestroy(PMEMRTREE prt);
|
|
extern PVOID _HeapRTreeGetLocked(PHEAPDATA phd, PMEMRTREE prt, UINT_PTR uiKey);
|
|
extern PVOID _HeapRTreeGet(PHEAPDATA phd, PMEMRTREE prt, UINT_PTR uiKey);
|
|
extern BOOL _HeapRTreeSet(PHEAPDATA phd, PMEMRTREE prt, UINT_PTR uiKey, PVOID pv);
|
|
|
|
CDECL_END
|
|
|
|
/*-------------------------------------
|
|
* Internal chunk management functions
|
|
*-------------------------------------
|
|
*/
|
|
|
|
/* Get chunk address for allocated address a. */
|
|
#define CHUNK_ADDR2BASE(phd, a) ALIGNMENT_ADDR2BASE((a), (phd)->uiChunkSizeMask)
|
|
|
|
/* Get chunk offset of allocated address a. */
|
|
#define CHUNK_ADDR2OFFSET(phd, a) ALIGNMENT_ADDR2OFFSET((a), (phd)->uiChunkSizeMask)
|
|
|
|
/* Return the smallest chunk size multiple that can contain a certain size. */
|
|
#define CHUNK_CEILING(phd, sz) ALIGNMENT_CEILING((sz), (phd)->uiChunkSizeMask)
|
|
|
|
CDECL_BEGIN
|
|
|
|
extern PVOID _HeapChunkAlloc(PHEAPDATA phd, SIZE_T sz, SIZE_T szAlignment, BOOL fBase, BOOL *pfZeroed);
|
|
extern void _HeapChunkUnmap(PHEAPDATA phd, PVOID pvChunk, SIZE_T sz);
|
|
extern void _HeapChunkDeAlloc(PHEAPDATA phd, PVOID pvChunk, SIZE_T sz, BOOL fUnmap);
|
|
extern HRESULT _HeapChunkSetup(PHEAPDATA phd);
|
|
extern void _HeapChunkShutdown(PHEAPDATA phd);
|
|
|
|
CDECL_END
|
|
|
|
/*------------------------------------
|
|
* Internal base management functions
|
|
*------------------------------------
|
|
*/
|
|
|
|
CDECL_BEGIN
|
|
|
|
extern PVOID _HeapBaseAlloc(PHEAPDATA phd, SIZE_T sz);
|
|
extern PEXTENT_NODE _HeapBaseNodeAlloc(PHEAPDATA phd);
|
|
extern void _HeapBaseNodeDeAlloc(PHEAPDATA phd, PEXTENT_NODE pexn);
|
|
extern HRESULT _HeapBaseSetup(PHEAPDATA phd);
|
|
extern void _HeapBaseShutdown(PHEAPDATA phd);
|
|
|
|
CDECL_END
|
|
|
|
/*----------------------------
|
|
* Arena management functions
|
|
*----------------------------
|
|
*/
|
|
|
|
CDECL_BEGIN
|
|
|
|
extern HRESULT _HeapArenaSetup(PHEAPDATA phd);
|
|
extern void _HeapArenaShutdown(PHEAPDATA phd);
|
|
|
|
CDECL_END
|
|
|
|
#endif /* __ASM__ */
|
|
|
|
#endif /* __COMROGUE_INTERNALS__ */
|
|
|
|
#endif /* __HEAP_INTERNALS_H_INCLUDED */
|