new version of PagedPoolCacheAligned. Fixing the problem in subsys/win32k/objects/dc.c allows this code to work perfectly ( even when forcing all PagedPool allocations to be cache aligned. Also implemented buffer over/under-run detection for the PagedPool.

svn path=/trunk/; revision=5402
This commit is contained in:
Royce Mitchell III 2003-08-04 00:39:58 +00:00
parent 330f6dce4f
commit 78dffae46c
2 changed files with 208 additions and 16 deletions

View file

@ -26,5 +26,18 @@ extern ULONG MmPagedPoolSize;
*/
#define MM_KERNEL_MAP_SIZE (16*1024*1024)
/*
* (number of bytes-1) per cache page
*
* FIXME - different architectures have different cache alignments...
*/
#define MM_CACHE_ALIGN_BYTES 31
#define MM_CACHE_ALIGN_UP(ptr) (PVOID)((((size_t)(ptr) + MM_CACHE_ALIGN_BYTES)) & (~MM_CACHE_ALIGN_BYTES))
#define MM_CACHE_ALIGN_DOWN(ptr) (PVOID)(((size_t)(ptr)) & (~MM_CACHE_ALIGN_BYTES))
#define MM_IS_CACHE_ALIGNED(ptr) ((PVOID)(ptr) == MM_CACHE_ALIGN(ptr))
#endif /* __INTERNAL_POOL_H */

View file

@ -1,4 +1,4 @@
/* $Id: ppool.c,v 1.14 2003/07/29 18:51:11 royce Exp $
/* $Id: ppool.c,v 1.15 2003/08/04 00:39:58 royce Exp $
*
* COPYRIGHT: See COPYING in the top level directory
* PROJECT: ReactOS kernel
@ -19,15 +19,26 @@
/* GLOBALS *******************************************************************/
#undef assert
#define assert(x) if (!(x)) {DbgPrint("Assertion "#x" failed at %s:%d\n", __FILE__,__LINE__); KeBugCheck(0); }
#define ASSERT_SIZE(n) assert ( (n) <= MmPagedPoolSize && (n) >= 0 )
#define ASSERT_PTR(p) assert ( ((size_t)(p)) >= ((size_t)MmPagedPoolBase) && ((size_t)(p)) < ((size_t)(MmPagedPoolBase+MmPagedPoolSize)) )
// to disable buffer over/under-run detection, set the following macro to 0
#define MM_PPOOL_BOUNDARY_BYTES 4
typedef struct _MM_PPOOL_FREE_BLOCK_HEADER
{
ULONG Size;
LONG Size;
struct _MM_PPOOL_FREE_BLOCK_HEADER* NextFree;
} MM_PPOOL_FREE_BLOCK_HEADER, *PMM_PPOOL_FREE_BLOCK_HEADER;
typedef struct _MM_PPOOL_USED_BLOCK_HEADER
{
ULONG Size;
LONG Size;
#if MM_PPOOL_BOUNDARY_BYTES
LONG UserSize; // how many bytes the user actually asked for...
#endif//MM_PPOOL_BOUNDARY_BYTES
} MM_PPOOL_USED_BLOCK_HEADER, *PMM_PPOOL_USED_BLOCK_HEADER;
PVOID MmPagedPoolBase;
@ -37,20 +48,19 @@ static PMM_PPOOL_FREE_BLOCK_HEADER MmPagedPoolFirstFreeBlock;
/* FUNCTIONS *****************************************************************/
inline static void* block_to_address (
MM_PPOOL_USED_BLOCK_HEADER* blk )
inline static void* block_to_address ( PVOID blk )
/*
* FUNCTION: Translate a block header address to the corresponding block
* address (internal)
*/
{
return ( (void *) ((char*)blk + sizeof(MM_PPOOL_USED_BLOCK_HEADER)) );
return ( (void *) ((char*)blk + sizeof(MM_PPOOL_USED_BLOCK_HEADER) + MM_PPOOL_BOUNDARY_BYTES) );
}
inline static MM_PPOOL_USED_BLOCK_HEADER* address_to_block(void* addr)
inline static PMM_PPOOL_USED_BLOCK_HEADER address_to_block(PVOID addr)
{
return (MM_PPOOL_USED_BLOCK_HEADER *)
( ((char*)addr) - sizeof(MM_PPOOL_USED_BLOCK_HEADER) );
return (PMM_PPOOL_USED_BLOCK_HEADER)
( ((char*)addr) - sizeof(MM_PPOOL_USED_BLOCK_HEADER) - MM_PPOOL_BOUNDARY_BYTES );
}
VOID MmInitializePagedPool(VOID)
@ -67,6 +77,31 @@ VOID MmInitializePagedPool(VOID)
ExInitializeFastMutex(&MmPagedPoolLock);
}
#if 1
static void VerifyPagedPool ( int line )
{
PMM_PPOOL_FREE_BLOCK_HEADER p = MmPagedPoolFirstFreeBlock;
int count = 0;
//DbgPrint ( "VerifyPagedPool(%i):\n", line );
while ( p )
{
//DbgPrint ( " 0x%x: %lu bytes (next 0x%x)\n", p, p->Size, p->NextFree );
ASSERT_PTR(p);
ASSERT_SIZE(p->Size);
count++;
p = p->NextFree;
}
//DbgPrint ( "VerifyPagedPool(%i): (%lu blocks)\n", line, count );
}
#define VerifyPagedPool() VerifyPagedPool(__LINE__)
#undef DPRINT
#define DPRINT(...)
#else
#define VerifyPagedPool()
#undef DPRINT
#define DPRINT(...)
#endif
/**********************************************************************
* NAME INTERNAL
* ExAllocatePagedPoolWithTag@12
@ -99,10 +134,13 @@ ExAllocatePagedPoolWithTag (IN POOL_TYPE PoolType,
return(NULL);
}
DPRINT ( "ExAllocatePagedPoolWithTag(%i,%lu,%lu)\n", PoolType, NumberOfBytes, Tag );
VerifyPagedPool();
/*
* Calculate the total number of bytes we will need.
*/
BlockSize = NumberOfBytes + sizeof(MM_PPOOL_USED_BLOCK_HEADER);
BlockSize = NumberOfBytes + sizeof(MM_PPOOL_USED_BLOCK_HEADER) + 2*MM_PPOOL_BOUNDARY_BYTES;
if (BlockSize < sizeof(MM_PPOOL_FREE_BLOCK_HEADER))
{
/* At least we need the size of the free block header. */
@ -112,16 +150,90 @@ ExAllocatePagedPoolWithTag (IN POOL_TYPE PoolType,
ExAcquireFastMutex(&MmPagedPoolLock);
/*
* Find the best fitting block.
* Find the best-fitting block.
*/
PreviousBlock = NULL;
BestPreviousBlock = BestBlock = NULL;
CurrentBlock = MmPagedPoolFirstFreeBlock;
if ( PoolType == PagedPoolCacheAligned )
{
PVOID BestAlignedAddr = NULL;
while ( CurrentBlock != NULL )
{
if (CurrentBlock->Size >= BlockSize &&
(BestBlock == NULL ||
(BestBlock->Size - BlockSize) > (CurrentBlock->Size - BlockSize)))
PVOID Addr = block_to_address(CurrentBlock);
PVOID CurrentBlockEnd = Addr + CurrentBlock->Size;
/* calculate last cache-aligned address available within this block */
PVOID AlignedAddr = MM_CACHE_ALIGN_DOWN(CurrentBlockEnd-NumberOfBytes-MM_PPOOL_BOUNDARY_BYTES);
assert ( AlignedAddr+NumberOfBytes+MM_PPOOL_BOUNDARY_BYTES <= CurrentBlockEnd );
/* special case, this address is already cache-aligned, and the right size */
if ( Addr == AlignedAddr )
{
BestAlignedAddr = AlignedAddr;
BestPreviousBlock = PreviousBlock;
BestBlock = CurrentBlock;
break;
}
else if ( Addr < (PVOID)address_to_block(AlignedAddr) )
{
/*
* there's enough room to allocate our cache-aligned memory out
* of this block, see if it's a better choice than any previous
* finds
*/
if ( BestBlock == NULL || BestBlock->Size > CurrentBlock->Size )
{
BestAlignedAddr = AlignedAddr;
BestPreviousBlock = PreviousBlock;
BestBlock = CurrentBlock;
}
}
PreviousBlock = CurrentBlock;
CurrentBlock = CurrentBlock->NextFree;
}
/*
* we found a best block can/should we chop a few bytes off the beginning
* into a separate memory block?
*/
if ( BestBlock != NULL )
{
PVOID Addr = block_to_address(BestBlock);
if ( BestAlignedAddr != Addr )
{
PMM_PPOOL_FREE_BLOCK_HEADER NewFreeBlock =
(PMM_PPOOL_FREE_BLOCK_HEADER)address_to_block(BestAlignedAddr);
assert ( BestAlignedAddr > Addr );
NewFreeBlock->Size = Addr + BestBlock->Size - BestAlignedAddr;
ASSERT_SIZE(NewFreeBlock->Size);
BestBlock->Size = (size_t)NewFreeBlock - (size_t)Addr;
ASSERT_SIZE(BestBlock->Size);
DPRINT ( "breaking off preceding bytes into their own block...\n" );
DPRINT ( "NewFreeBlock 0x%x Size %lu (Old Block's new size %lu) NextFree 0x%x\n",
NewFreeBlock, NewFreeBlock->Size, BestBlock->Size, BestBlock->NextFree );
/* insert the new block into the chain */
NewFreeBlock->NextFree = BestBlock->NextFree;
BestBlock->NextFree = NewFreeBlock;
/* we want the following code to use our cache-aligned block */
BestPreviousBlock = BestBlock;
BestBlock = NewFreeBlock;
//VerifyPagedPool();
}
}
}
/*
* non-cache-aligned block search
*/
else while ( CurrentBlock != NULL )
{
if ( CurrentBlock->Size >= BlockSize
&& ( BestBlock == NULL || BestBlock->Size > CurrentBlock->Size )
)
{
BestPreviousBlock = PreviousBlock;
BestBlock = CurrentBlock;
@ -136,41 +248,66 @@ ExAllocatePagedPoolWithTag (IN POOL_TYPE PoolType,
*/
if (BestBlock == NULL)
{
DPRINT("ExAllocatePagedPoolWithTag() - nothing suitable found, returning NULL\n" );
ExReleaseFastMutex(&MmPagedPoolLock);
return(NULL);
}
DPRINT("BestBlock 0x%x NextFree 0x%x\n", BestBlock, BestBlock->NextFree );
//VerifyPagedPool();
/*
* Is there enough space to create a second block from the unused portion.
*/
if ((BestBlock->Size - BlockSize) > sizeof(MM_PPOOL_FREE_BLOCK_HEADER))
if ( BestBlock->Size > BlockSize
&& (BestBlock->Size - BlockSize) > sizeof(MM_PPOOL_FREE_BLOCK_HEADER)
)
{
ULONG NewSize = BestBlock->Size - BlockSize;
ASSERT_SIZE ( NewSize );
//DPRINT("creating 2nd block from unused portion\n");
DPRINT("BestBlock 0x%x Size 0x%x BlockSize 0x%x NewSize 0x%x\n",
BestBlock, BestBlock->Size, BlockSize, NewSize );
/*
* Create the new free block.
*/
//DPRINT("creating the new free block");
NextBlock = (PMM_PPOOL_FREE_BLOCK_HEADER)((char*)BestBlock + BlockSize);
//DPRINT(".");
NextBlock->Size = NewSize;
ASSERT_SIZE ( NextBlock->Size );
//DPRINT(".");
NextBlock->NextFree = BestBlock->NextFree;
//DPRINT(".\n");
/*
* Replace the old free block with it.
*/
//DPRINT("replacing old free block with it");
if (BestPreviousBlock == NULL)
{
//DPRINT("(from beginning)");
MmPagedPoolFirstFreeBlock = NextBlock;
}
else
{
//DPRINT("(from previous)");
BestPreviousBlock->NextFree = NextBlock;
}
//DPRINT(".\n");
/*
* Create the new used block header.
*/
//DPRINT("create new used block header");
NewBlock = (PMM_PPOOL_USED_BLOCK_HEADER)BestBlock;
//DPRINT(".");
NewBlock->Size = BlockSize;
ASSERT_SIZE ( NewBlock->Size );
//DPRINT(".\n");
}
else
{
@ -179,6 +316,7 @@ ExAllocatePagedPoolWithTag (IN POOL_TYPE PoolType,
/*
* Remove the selected block from the list of free blocks.
*/
//DPRINT ( "Removing selected block from free block list\n" );
if (BestPreviousBlock == NULL)
{
MmPagedPoolFirstFreeBlock = BestBlock->NextFree;
@ -193,14 +331,35 @@ ExAllocatePagedPoolWithTag (IN POOL_TYPE PoolType,
*/
NewBlock = (PMM_PPOOL_USED_BLOCK_HEADER)BestBlock;
NewBlock->Size = NewSize;
ASSERT_SIZE ( NewBlock->Size );
}
VerifyPagedPool();
ExReleaseFastMutex(&MmPagedPoolLock);
BlockAddress = block_to_address ( NewBlock );
memset(BlockAddress, 0, NumberOfBytes);
#if MM_PPOOL_BOUNDARY_BYTES
NewBlock->UserSize = NumberOfBytes;
// write out buffer-overrun detection bytes
{
int i;
PUCHAR Addr = (PUCHAR)BlockAddress;
//DbgPrint ( "writing buffer-overrun detection bytes" );
for ( i = 0; i < MM_PPOOL_BOUNDARY_BYTES; i++ )
{
//DbgPrint(".");
*(Addr-i-1) = 0xCD;
//DbgPrint("o");
*(Addr+NewBlock->UserSize+i) = 0xCD;
}
//DbgPrint ( "done!\n" );
}
#endif//MM_PPOOL_BOUNDARY_BYTES
return(BlockAddress);
}
@ -215,12 +374,28 @@ ExFreePagedPool(IN PVOID Block)
PMM_PPOOL_FREE_BLOCK_HEADER NextBlock;
PMM_PPOOL_FREE_BLOCK_HEADER NextNextBlock;
#if MM_PPOOL_BOUNDARY_BYTES
// write out buffer-overrun detection bytes
{
int i;
PUCHAR Addr = (PUCHAR)Block;
//DbgPrint ( "checking buffer-overrun detection bytes..." );
for ( i = 0; i < MM_PPOOL_BOUNDARY_BYTES; i++ )
{
assert ( *(Addr-i-1) == 0xCD );
assert ( *(Addr+UsedBlock->UserSize+i) == 0xCD );
}
//DbgPrint ( "done!\n" );
}
#endif//MM_PPOOL_BOUNDARY_BYTES
ExAcquireFastMutex(&MmPagedPoolLock);
/*
* Begin setting up the newly freed block's header.
*/
FreeBlock->Size = UsedSize;
ASSERT_SIZE ( FreeBlock->Size );
/*
* Find the blocks immediately before and after the newly freed block on the free list.
@ -255,6 +430,7 @@ ExFreePagedPool(IN PVOID Block)
((char*)FreeBlock + FreeBlock->Size) == (char*)NextBlock)
{
FreeBlock->Size = FreeBlock->Size + NextBlock->Size;
ASSERT_SIZE ( FreeBlock->Size );
FreeBlock->NextFree = NextBlock->NextFree;
NextNextBlock = NextBlock->NextFree;
}
@ -271,9 +447,12 @@ ExFreePagedPool(IN PVOID Block)
((char*)PreviousBlock + PreviousBlock->Size) == (char*)FreeBlock)
{
PreviousBlock->Size = PreviousBlock->Size + FreeBlock->Size;
ASSERT_SIZE ( PreviousBlock->Size );
PreviousBlock->NextFree = NextNextBlock;
}
VerifyPagedPool();
ExReleaseFastMutex(&MmPagedPoolLock);
}