reactos/drivers/filesystems/fastfat_new/allocsup.c
2020-12-22 11:08:21 +01:00

5293 lines
157 KiB
C

/*++
Copyright (c) 1990-2000 Microsoft Corporation
Module Name:
AllocSup.c
Abstract:
This module implements the Allocation support routines for Fat.
--*/
#include "fatprocs.h"
//
// The Bug check file id for this module
//
#define BugCheckFileId (FAT_BUG_CHECK_ALLOCSUP)
//
// Local debug trace level
//
#define Dbg (DEBUG_TRACE_ALLOCSUP)
#define FatMin(a, b) ((a) < (b) ? (a) : (b))
//
// Define prefetch page count for the FAT
//
#define FAT_PREFETCH_PAGE_COUNT 0x100
//
// Local support routine prototypes
//
VOID
FatLookupFatEntry(
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG FatIndex,
IN OUT PULONG FatEntry,
IN OUT PFAT_ENUMERATION_CONTEXT Context
);
VOID
FatSetFatRun(
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG StartingFatIndex,
IN ULONG ClusterCount,
IN BOOLEAN ChainTogether
);
UCHAR
FatLogOf(
IN ULONG Value
);
//
// Note that the KdPrint below will ONLY fire when the assert does. Leave it
// alone.
//
#if DBG
#define ASSERT_CURRENT_WINDOW_GOOD(VCB) { \
ULONG FreeClusterBitMapClear; \
NT_ASSERT( (VCB)->FreeClusterBitMap.Buffer != NULL ); \
FreeClusterBitMapClear = RtlNumberOfClearBits(&(VCB)->FreeClusterBitMap); \
if ((VCB)->CurrentWindow->ClustersFree != FreeClusterBitMapClear) { \
KdPrint(("FAT: ClustersFree %x h != FreeClusterBitMapClear %x h\n", \
(VCB)->CurrentWindow->ClustersFree, \
FreeClusterBitMapClear)); \
} \
NT_ASSERT( (VCB)->CurrentWindow->ClustersFree == FreeClusterBitMapClear ); \
}
#else
#define ASSERT_CURRENT_WINDOW_GOOD(VCB)
#endif
//
// The following macros provide a convenient way of hiding the details
// of bitmap allocation schemes.
//
//
// VOID
// FatLockFreeClusterBitMap (
// IN PVCB Vcb
// );
//
#define FatLockFreeClusterBitMap(VCB) { \
NT_ASSERT(KeAreApcsDisabled()); \
ExAcquireFastMutexUnsafe( &(VCB)->FreeClusterBitMapMutex ); \
ASSERT_CURRENT_WINDOW_GOOD(VCB) \
}
//
// VOID
// FatUnlockFreeClusterBitMap (
// IN PVCB Vcb
// );
//
#define FatUnlockFreeClusterBitMap(VCB) { \
ASSERT_CURRENT_WINDOW_GOOD(VCB) \
NT_ASSERT(KeAreApcsDisabled()); \
ExReleaseFastMutexUnsafe( &(VCB)->FreeClusterBitMapMutex ); \
}
//
// BOOLEAN
// FatIsClusterFree (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG FatIndex
// );
//
#define FatIsClusterFree(IRPCONTEXT,VCB,FAT_INDEX) \
(RtlCheckBit(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2) == 0)
//
// VOID
// FatFreeClusters (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG FatIndex,
// IN ULONG ClusterCount
// );
//
#define FatFreeClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
if ((CLUSTER_COUNT) == 1) { \
FatSetFatEntry((IRPCONTEXT),(VCB),(FAT_INDEX),FAT_CLUSTER_AVAILABLE); \
} else { \
FatSetFatRun((IRPCONTEXT),(VCB),(FAT_INDEX),(CLUSTER_COUNT),FALSE); \
} \
}
//
// VOID
// FatAllocateClusters (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG FatIndex,
// IN ULONG ClusterCount
// );
//
#define FatAllocateClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
if ((CLUSTER_COUNT) == 1) { \
FatSetFatEntry((IRPCONTEXT),(VCB),(FAT_INDEX),FAT_CLUSTER_LAST); \
} else { \
FatSetFatRun((IRPCONTEXT),(VCB),(FAT_INDEX),(CLUSTER_COUNT),TRUE); \
} \
}
//
// VOID
// FatUnreserveClusters (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG FatIndex,
// IN ULONG ClusterCount
// );
//
#define FatUnreserveClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
NT_ASSERT( (FAT_INDEX) + (CLUSTER_COUNT) - 2 <= (VCB)->FreeClusterBitMap.SizeOfBitMap );\
NT_ASSERT( (FAT_INDEX) >= 2); \
RtlClearBits(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2,(CLUSTER_COUNT)); \
if ((FAT_INDEX) < (VCB)->ClusterHint) { \
(VCB)->ClusterHint = (FAT_INDEX); \
} \
}
//
// VOID
// FatReserveClusters (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG FatIndex,
// IN ULONG ClusterCount
// );
//
// Handle wrapping the hint back to the front.
//
#define FatReserveClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
ULONG _AfterRun = (FAT_INDEX) + (CLUSTER_COUNT); \
NT_ASSERT( (FAT_INDEX) + (CLUSTER_COUNT) - 2 <= (VCB)->FreeClusterBitMap.SizeOfBitMap );\
NT_ASSERT( (FAT_INDEX) >= 2); \
RtlSetBits(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2,(CLUSTER_COUNT)); \
\
if (_AfterRun - 2 >= (VCB)->FreeClusterBitMap.SizeOfBitMap) { \
_AfterRun = 2; \
} \
if (RtlCheckBit(&(VCB)->FreeClusterBitMap, _AfterRun - 2)) { \
(VCB)->ClusterHint = RtlFindClearBits( &(VCB)->FreeClusterBitMap, 1, _AfterRun - 2) + 2; \
if (1 == (VCB)->ClusterHint) { \
(VCB)->ClusterHint = 2; \
} \
} \
else { \
(VCB)->ClusterHint = _AfterRun; \
} \
}
//
// ULONG
// FatFindFreeClusterRun (
// IN PIRP_CONTEXT IrpContext,
// IN PVCB Vcb,
// IN ULONG ClusterCount,
// IN ULONG AlternateClusterHint
// );
//
// Do a special check if only one cluster is desired.
//
#define FatFindFreeClusterRun(IRPCONTEXT,VCB,CLUSTER_COUNT,CLUSTER_HINT) ( \
(CLUSTER_COUNT == 1) && \
FatIsClusterFree((IRPCONTEXT), (VCB), (CLUSTER_HINT)) ? \
(CLUSTER_HINT) : \
RtlFindClearBits( &(VCB)->FreeClusterBitMap, \
(CLUSTER_COUNT), \
(CLUSTER_HINT) - 2) + 2 \
)
//
// FAT32: Define the maximum size of the FreeClusterBitMap to be the
// maximum size of a FAT16 FAT. If there are more clusters on the
// volume than can be represented by this many bytes of bitmap, the
// FAT will be split into "buckets", each of which does fit.
//
// Note this count is in clusters/bits of bitmap.
//
#define MAX_CLUSTER_BITMAP_SIZE (1 << 16)
//
// Calculate the window a given cluster number is in.
//
#define FatWindowOfCluster(C) (((C) - 2) / MAX_CLUSTER_BITMAP_SIZE)
#ifdef ALLOC_PRAGMA
#pragma alloc_text(PAGE, FatAddFileAllocation)
#pragma alloc_text(PAGE, FatAllocateDiskSpace)
#pragma alloc_text(PAGE, FatDeallocateDiskSpace)
#pragma alloc_text(PAGE, FatExamineFatEntries)
#pragma alloc_text(PAGE, FatInterpretClusterType)
#pragma alloc_text(PAGE, FatLogOf)
#pragma alloc_text(PAGE, FatLookupFatEntry)
#pragma alloc_text(PAGE, FatLookupFileAllocation)
#pragma alloc_text(PAGE, FatLookupFileAllocationSize)
#pragma alloc_text(PAGE, FatMergeAllocation)
#pragma alloc_text(PAGE, FatSetFatEntry)
#pragma alloc_text(PAGE, FatSetFatRun)
#pragma alloc_text(PAGE, FatSetupAllocationSupport)
#pragma alloc_text(PAGE, FatSplitAllocation)
#pragma alloc_text(PAGE, FatTearDownAllocationSupport)
#pragma alloc_text(PAGE, FatTruncateFileAllocation)
#endif
#ifdef __REACTOS__
static
#endif
INLINE
ULONG
FatSelectBestWindow(
IN PVCB Vcb
)
/*++
Routine Description:
Choose a window to allocate clusters from. Order of preference is:
1. First window with >50% free clusters
2. First empty window
3. Window with greatest number of free clusters.
Arguments:
Vcb - Supplies the Vcb for the volume
Return Value:
'Best window' number (index into Vcb->Windows[])
--*/
{
ULONG i, Fave = 0;
ULONG MaxFree = 0;
ULONG FirstEmpty = (ULONG)-1;
ULONG ClustersPerWindow = MAX_CLUSTER_BITMAP_SIZE;
NT_ASSERT( 1 != Vcb->NumberOfWindows);
for (i = 0; i < Vcb->NumberOfWindows; i++) {
if (Vcb->Windows[i].ClustersFree == ClustersPerWindow) {
if (-1 == FirstEmpty) {
//
// Keep note of the first empty window on the disc
//
FirstEmpty = i;
}
}
else if (Vcb->Windows[i].ClustersFree > MaxFree) {
//
// This window has the most free clusters, so far
//
MaxFree = Vcb->Windows[i].ClustersFree;
Fave = i;
//
// If this window has >50% free clusters, then we will take it,
// so don't bother considering more windows.
//
if (MaxFree >= (ClustersPerWindow >> 1)) {
break;
}
}
}
//
// If there were no windows with 50% or more freespace, then select the
// first empty window on the disc, if any - otherwise we'll just go with
// the one with the most free clusters.
//
if ((MaxFree < (ClustersPerWindow >> 1)) && (-1 != FirstEmpty)) {
Fave = FirstEmpty;
}
return Fave;
}
VOID
FatSetupAllocationSupport (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb
)
/*++
Routine Description:
This routine fills in the Allocation Support structure in the Vcb.
Most entries are computed using fat.h macros supplied with data from
the Bios Parameter Block. The free cluster count, however, requires
going to the Fat and actually counting free sectors. At the same time
the free cluster bit map is initalized.
Arguments:
Vcb - Supplies the Vcb to fill in.
--*/
{
ULONG BitIndex;
ULONG ClustersDescribableByFat;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatSetupAllocationSupport\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
//
// Compute a number of fields for Vcb.AllocationSupport
//
Vcb->AllocationSupport.RootDirectoryLbo = FatRootDirectoryLbo( &Vcb->Bpb );
Vcb->AllocationSupport.RootDirectorySize = FatRootDirectorySize( &Vcb->Bpb );
Vcb->AllocationSupport.FileAreaLbo = FatFileAreaLbo( &Vcb->Bpb );
Vcb->AllocationSupport.NumberOfClusters = FatNumberOfClusters( &Vcb->Bpb );
Vcb->AllocationSupport.FatIndexBitSize = FatIndexBitSize( &Vcb->Bpb );
Vcb->AllocationSupport.LogOfBytesPerSector = FatLogOf(Vcb->Bpb.BytesPerSector);
Vcb->AllocationSupport.LogOfBytesPerCluster = FatLogOf(FatBytesPerCluster( &Vcb->Bpb ));
Vcb->AllocationSupport.NumberOfFreeClusters = 0;
//
// Deal with a bug in DOS 5 format, if the Fat is not big enough to
// describe all the clusters on the disk, reduce this number. We expect
// that fat32 volumes will not have this problem.
//
// Turns out this was not a good assumption. We have to do this always now.
//
ClustersDescribableByFat = ( ((FatIsFat32(Vcb)? Vcb->Bpb.LargeSectorsPerFat :
Vcb->Bpb.SectorsPerFat) *
Vcb->Bpb.BytesPerSector * 8)
/ FatIndexBitSize(&Vcb->Bpb) ) - 2;
if (Vcb->AllocationSupport.NumberOfClusters > ClustersDescribableByFat) {
Vcb->AllocationSupport.NumberOfClusters = ClustersDescribableByFat;
}
//
// Extend the virtual volume file to include the Fat
//
{
CC_FILE_SIZES FileSizes;
FileSizes.AllocationSize.QuadPart =
FileSizes.FileSize.QuadPart = (FatReservedBytes( &Vcb->Bpb ) +
FatBytesPerFat( &Vcb->Bpb ));
FileSizes.ValidDataLength = FatMaxLarge;
if ( Vcb->VirtualVolumeFile->PrivateCacheMap == NULL ) {
FatInitializeCacheMap( Vcb->VirtualVolumeFile,
&FileSizes,
TRUE,
&FatData.CacheManagerNoOpCallbacks,
Vcb );
} else {
CcSetFileSizes( Vcb->VirtualVolumeFile, &FileSizes );
}
}
_SEH2_TRY {
if (FatIsFat32(Vcb) &&
Vcb->AllocationSupport.NumberOfClusters > MAX_CLUSTER_BITMAP_SIZE) {
Vcb->NumberOfWindows = (Vcb->AllocationSupport.NumberOfClusters +
MAX_CLUSTER_BITMAP_SIZE - 1) /
MAX_CLUSTER_BITMAP_SIZE;
} else {
Vcb->NumberOfWindows = 1;
}
Vcb->Windows = FsRtlAllocatePoolWithTag( PagedPool,
Vcb->NumberOfWindows * sizeof(FAT_WINDOW),
TAG_FAT_WINDOW );
RtlInitializeBitMap( &Vcb->FreeClusterBitMap,
NULL,
0 );
//
// Chose a FAT window to begin operation in.
//
if (Vcb->NumberOfWindows > 1) {
//
// Read the fat and count up free clusters. We bias by the two reserved
// entries in the FAT.
//
FatExamineFatEntries( IrpContext, Vcb,
2,
Vcb->AllocationSupport.NumberOfClusters + 2 - 1,
TRUE,
NULL,
NULL);
//
// Pick a window to begin allocating from
//
Vcb->CurrentWindow = &Vcb->Windows[ FatSelectBestWindow( Vcb)];
} else {
Vcb->CurrentWindow = &Vcb->Windows[0];
//
// Carefully bias ourselves by the two reserved entries in the FAT.
//
Vcb->CurrentWindow->FirstCluster = 2;
Vcb->CurrentWindow->LastCluster = Vcb->AllocationSupport.NumberOfClusters + 2 - 1;
}
//
// Now transition to the FAT window we have chosen.
//
FatExamineFatEntries( IrpContext, Vcb,
0,
0,
FALSE,
Vcb->CurrentWindow,
NULL);
//
// Now set the ClusterHint to the first free bit in our favorite
// window (except the ClusterHint is off by two).
//
Vcb->ClusterHint =
(BitIndex = RtlFindClearBits( &Vcb->FreeClusterBitMap, 1, 0 )) != -1 ?
BitIndex + 2 : 2;
} _SEH2_FINALLY {
DebugUnwind( FatSetupAllocationSupport );
//
// If we hit an exception, back out.
//
if (_SEH2_AbnormalTermination()) {
FatTearDownAllocationSupport( IrpContext, Vcb );
}
} _SEH2_END;
return;
}
VOID
FatTearDownAllocationSupport (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb
)
/*++
Routine Description:
This routine prepares the volume for closing. Specifically, we must
release the free fat bit map buffer, and uninitialize the dirty fat
Mcb.
Arguments:
Vcb - Supplies the Vcb to fill in.
Return Value:
VOID
--*/
{
DebugTrace(+1, Dbg, "FatTearDownAllocationSupport\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
PAGED_CODE();
//
// If there are FAT buckets, free them.
//
if ( Vcb->Windows != NULL ) {
ExFreePool( Vcb->Windows );
Vcb->Windows = NULL;
}
//
// Free the memory associated with the free cluster bitmap.
//
if ( Vcb->FreeClusterBitMap.Buffer != NULL ) {
ExFreePool( Vcb->FreeClusterBitMap.Buffer );
//
// NULL this field as an flag.
//
Vcb->FreeClusterBitMap.Buffer = NULL;
}
//
// And remove all the runs in the dirty fat Mcb
//
FatRemoveMcbEntry( Vcb, &Vcb->DirtyFatMcb, 0, 0xFFFFFFFF );
DebugTrace(-1, Dbg, "FatTearDownAllocationSupport -> (VOID)\n", 0);
UNREFERENCED_PARAMETER( IrpContext );
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatLookupFileAllocation (
IN PIRP_CONTEXT IrpContext,
IN PFCB FcbOrDcb,
IN VBO Vbo,
OUT PLBO Lbo,
OUT PULONG ByteCount,
OUT PBOOLEAN Allocated,
OUT PBOOLEAN EndOnMax,
OUT PULONG Index
)
/*++
Routine Description:
This routine looks up the existing mapping of VBO to LBO for a
file/directory. The information it queries is either stored in the
mcb field of the fcb/dcb or it is stored on in the fat table and
needs to be retrieved and decoded, and updated in the mcb.
Arguments:
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being queried
Vbo - Supplies the VBO whose LBO we want returned
Lbo - Receives the LBO corresponding to the input Vbo if one exists
ByteCount - Receives the number of bytes within the run the run
that correpond between the input vbo and output lbo.
Allocated - Receives TRUE if the Vbo does have a corresponding Lbo
and FALSE otherwise.
EndOnMax - Receives TRUE if the run ends in the maximal FAT cluster,
which results in a fractional bytecount.
Index - Receives the Index of the run
--*/
{
VBO CurrentVbo;
LBO CurrentLbo;
LBO PriorLbo;
VBO FirstVboOfCurrentRun = 0;
LBO FirstLboOfCurrentRun;
BOOLEAN LastCluster;
ULONG Runs;
PVCB Vcb;
FAT_ENTRY FatEntry;
ULONG BytesPerCluster;
ULARGE_INTEGER BytesOnVolume;
FAT_ENUMERATION_CONTEXT Context;
PAGED_CODE();
Vcb = FcbOrDcb->Vcb;
DebugTrace(+1, Dbg, "FatLookupFileAllocation\n", 0);
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
DebugTrace( 0, Dbg, " Vbo = %8lx\n", Vbo);
DebugTrace( 0, Dbg, " pLbo = %8lx\n", Lbo);
DebugTrace( 0, Dbg, " pByteCount = %8lx\n", ByteCount);
DebugTrace( 0, Dbg, " pAllocated = %8lx\n", Allocated);
Context.Bcb = NULL;
*EndOnMax = FALSE;
//
// Check the trivial case that the mapping is already in our
// Mcb.
//
if ( FatLookupMcbEntry(Vcb, &FcbOrDcb->Mcb, Vbo, Lbo, ByteCount, Index) ) {
*Allocated = TRUE;
NT_ASSERT( *ByteCount != 0 );
//
// Detect the overflow case, trim and claim the condition.
//
if (Vbo + *ByteCount == 0) {
*EndOnMax = TRUE;
}
DebugTrace( 0, Dbg, "Found run in Mcb.\n", 0);
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
return;
}
//
// Initialize the Vcb, the cluster size, LastCluster, and
// FirstLboOfCurrentRun (to be used as an indication of the first
// iteration through the following while loop).
//
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
BytesOnVolume.QuadPart = UInt32x32To64( Vcb->AllocationSupport.NumberOfClusters, BytesPerCluster );
LastCluster = FALSE;
FirstLboOfCurrentRun = 0;
//
// Discard the case that the request extends beyond the end of
// allocation. Note that if the allocation size if not known
// AllocationSize is set to 0xffffffff.
//
if ( Vbo >= FcbOrDcb->Header.AllocationSize.LowPart ) {
*Allocated = FALSE;
DebugTrace( 0, Dbg, "Vbo beyond end of file.\n", 0);
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
return;
}
//
// The Vbo is beyond the last Mcb entry. So we adjust Current Vbo/Lbo
// and FatEntry to describe the beginning of the last entry in the Mcb.
// This is used as initialization for the following loop.
//
// If the Mcb was empty, we start at the beginning of the file with
// CurrentVbo set to 0 to indicate a new run.
//
if (FatLookupLastMcbEntry( Vcb, &FcbOrDcb->Mcb, &CurrentVbo, &CurrentLbo, &Runs )) {
DebugTrace( 0, Dbg, "Current Mcb size = %8lx.\n", CurrentVbo + 1);
CurrentVbo -= (BytesPerCluster - 1);
CurrentLbo -= (BytesPerCluster - 1);
//
// Convert an index to a count.
//
Runs += 1;
} else {
DebugTrace( 0, Dbg, "Mcb empty.\n", 0);
//
// Check for an FcbOrDcb that has no allocation
//
if (FcbOrDcb->FirstClusterOfFile == 0) {
*Allocated = FALSE;
DebugTrace( 0, Dbg, "File has no allocation.\n", 0);
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
return;
} else {
CurrentVbo = 0;
CurrentLbo = FatGetLboFromIndex( Vcb, FcbOrDcb->FirstClusterOfFile );
FirstVboOfCurrentRun = CurrentVbo;
FirstLboOfCurrentRun = CurrentLbo;
Runs = 0;
DebugTrace( 0, Dbg, "First Lbo of file = %8lx\n", CurrentLbo);
}
}
//
// Now we know that we are looking up a valid Vbo, but it is
// not in the Mcb, which is a monotonically increasing list of
// Vbo's. Thus we have to go to the Fat, and update
// the Mcb as we go. We use a try-finally to unpin the page
// of fat hanging around. Also we mark *Allocated = FALSE, so that
// the caller wont try to use the data if we hit an exception.
//
*Allocated = FALSE;
_SEH2_TRY {
FatEntry = (FAT_ENTRY)FatGetIndexFromLbo( Vcb, CurrentLbo );
//
// ASSERT that CurrentVbo and CurrentLbo are now cluster alligned.
// The assumption here, is that only whole clusters of Vbos and Lbos
// are mapped in the Mcb.
//
NT_ASSERT( ((CurrentLbo - Vcb->AllocationSupport.FileAreaLbo)
% BytesPerCluster == 0) &&
(CurrentVbo % BytesPerCluster == 0) );
//
// Starting from the first Vbo after the last Mcb entry, scan through
// the Fat looking for our Vbo. We continue through the Fat until we
// hit a noncontiguity beyond the desired Vbo, or the last cluster.
//
while ( !LastCluster ) {
//
// Get the next fat entry, and update our Current variables.
//
FatLookupFatEntry( IrpContext, Vcb, FatEntry, (PULONG)&FatEntry, &Context );
PriorLbo = CurrentLbo;
CurrentLbo = FatGetLboFromIndex( Vcb, FatEntry );
CurrentVbo += BytesPerCluster;
switch ( FatInterpretClusterType( Vcb, FatEntry )) {
//
// Check for a break in the Fat allocation chain.
//
case FatClusterAvailable:
case FatClusterReserved:
case FatClusterBad:
DebugTrace( 0, Dbg, "Break in allocation chain, entry = %d\n", FatEntry);
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> Fat Corrupt. Raise Status.\n", 0);
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
break;
//
// If this is the last cluster, we must update the Mcb and
// exit the loop.
//
case FatClusterLast:
//
// Assert we know where the current run started. If the
// Mcb was empty when we were called, thenFirstLboOfCurrentRun
// was set to the start of the file. If the Mcb contained an
// entry, then FirstLboOfCurrentRun was set on the first
// iteration through the loop. Thus if FirstLboOfCurrentRun
// is 0, then there was an Mcb entry and we are on our first
// iteration, meaing that the last cluster in the Mcb was
// really the last allocated cluster, but we checked Vbo
// against AllocationSize, and found it OK, thus AllocationSize
// must be too large.
//
// Note that, when we finally arrive here, CurrentVbo is actually
// the first Vbo beyond the file allocation and CurrentLbo is
// meaningless.
//
DebugTrace( 0, Dbg, "Read last cluster of file.\n", 0);
//
// Detect the case of the maximal file. Note that this really isn't
// a proper Vbo - those are zero-based, and this is a one-based number.
// The maximal file, of 2^32 - 1 bytes, has a maximum byte offset of
// 2^32 - 2.
//
// Just so we don't get confused here.
//
if (CurrentVbo == 0) {
*EndOnMax = TRUE;
CurrentVbo -= 1;
}
LastCluster = TRUE;
if (FirstLboOfCurrentRun != 0 ) {
DebugTrace( 0, Dbg, "Adding a run to the Mcb.\n", 0);
DebugTrace( 0, Dbg, " Vbo = %08lx.\n", FirstVboOfCurrentRun);
DebugTrace( 0, Dbg, " Lbo = %08lx.\n", FirstLboOfCurrentRun);
DebugTrace( 0, Dbg, " Length = %08lx.\n", CurrentVbo - FirstVboOfCurrentRun);
(VOID)FatAddMcbEntry( Vcb,
&FcbOrDcb->Mcb,
FirstVboOfCurrentRun,
FirstLboOfCurrentRun,
CurrentVbo - FirstVboOfCurrentRun );
Runs += 1;
}
//
// Being at the end of allocation, make sure we have found
// the Vbo. If we haven't, seeing as we checked VBO
// against AllocationSize, the real disk allocation is less
// than that of AllocationSize. This comes about when the
// real allocation is not yet known, and AllocaitonSize
// contains MAXULONG.
//
// KLUDGE! - If we were called by FatLookupFileAllocationSize
// Vbo is set to MAXULONG - 1, and AllocationSize to the lookup
// hint. Thus we merrily go along looking for a match that isn't
// there, but in the meantime building an Mcb. If this is
// the case, fill in AllocationSize and return.
//
if ( Vbo == MAXULONG - 1 ) {
*Allocated = FALSE;
FcbOrDcb->Header.AllocationSize.QuadPart = CurrentVbo;
DebugTrace( 0, Dbg, "New file allocation size = %08lx.\n", CurrentVbo);
try_return ( NOTHING );
}
//
// We will lie ever so slightly if we really terminated on the
// maximal byte of a file. It is really allocated.
//
if (Vbo >= CurrentVbo && !*EndOnMax) {
*Allocated = FALSE;
try_return ( NOTHING );
}
break;
//
// This is a continuation in the chain. If the run has a
// discontiguity at this point, update the Mcb, and if we are beyond
// the desired Vbo, this is the end of the run, so set LastCluster
// and exit the loop.
//
case FatClusterNext:
//
// This is the loop check. The Vbo must not be bigger than the size of
// the volume, and the Vbo must not have a) wrapped and b) not been at the
// very last cluster in the chain, for the case of the maximal file.
//
if ( CurrentVbo == 0 ||
(BytesOnVolume.HighPart == 0 && CurrentVbo > BytesOnVolume.LowPart)) {
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
}
if ( PriorLbo + BytesPerCluster != CurrentLbo ) {
//
// Note that on the first time through the loop
// (FirstLboOfCurrentRun == 0), we don't add the
// run to the Mcb since it curresponds to the last
// run already stored in the Mcb.
//
if ( FirstLboOfCurrentRun != 0 ) {
DebugTrace( 0, Dbg, "Adding a run to the Mcb.\n", 0);
DebugTrace( 0, Dbg, " Vbo = %08lx.\n", FirstVboOfCurrentRun);
DebugTrace( 0, Dbg, " Lbo = %08lx.\n", FirstLboOfCurrentRun);
DebugTrace( 0, Dbg, " Length = %08lx.\n", CurrentVbo - FirstVboOfCurrentRun);
FatAddMcbEntry( Vcb,
&FcbOrDcb->Mcb,
FirstVboOfCurrentRun,
FirstLboOfCurrentRun,
CurrentVbo - FirstVboOfCurrentRun );
Runs += 1;
}
//
// Since we are at a run boundry, with CurrentLbo and
// CurrentVbo being the first cluster of the next run,
// we see if the run we just added encompases the desired
// Vbo, and if so exit. Otherwise we set up two new
// First*boOfCurrentRun, and continue.
//
if (CurrentVbo > Vbo) {
LastCluster = TRUE;
} else {
FirstVboOfCurrentRun = CurrentVbo;
FirstLboOfCurrentRun = CurrentLbo;
}
}
break;
default:
DebugTrace(0, Dbg, "Illegal Cluster Type.\n", FatEntry);
#ifdef _MSC_VER
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
#endif
FatBugCheck( 0, 0, 0 );
break;
} // switch()
} // while()
//
// Load up the return parameters.
//
// On exit from the loop, Vbo still contains the desired Vbo, and
// CurrentVbo is the first byte after the run that contained the
// desired Vbo.
//
*Allocated = TRUE;
*Lbo = FirstLboOfCurrentRun + (Vbo - FirstVboOfCurrentRun);
*ByteCount = CurrentVbo - Vbo;
if (ARGUMENT_PRESENT(Index)) {
//
// Note that Runs only needs to be accurate with respect to where we
// ended. Since partial-lookup cases will occur without exclusive
// synchronization, the Mcb itself may be much bigger by now.
//
*Index = Runs - 1;
}
try_exit: NOTHING;
} _SEH2_FINALLY {
DebugUnwind( FatLookupFileAllocation );
//
// We are done reading the Fat, so unpin the last page of fat
// that is hanging around
//
FatUnpinBcb( IrpContext, Context.Bcb );
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
} _SEH2_END;
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatAddFileAllocation (
IN PIRP_CONTEXT IrpContext,
IN PFCB FcbOrDcb,
IN PFILE_OBJECT FileObject OPTIONAL,
IN ULONG DesiredAllocationSize
)
/*++
Routine Description:
This routine adds additional allocation to the specified file/directory.
Additional allocation is added by appending clusters to the file/directory.
If the file already has a sufficient allocation then this procedure
is effectively a noop.
Arguments:
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified.
This parameter must not specify the root dcb.
FileObject - If supplied inform the cache manager of the change.
DesiredAllocationSize - Supplies the minimum size, in bytes, that we want
allocated to the file/directory.
--*/
{
PVCB Vcb;
LARGE_MCB NewMcb = {0};
PLARGE_MCB McbToCleanup = NULL;
PDIRENT Dirent = NULL;
ULONG NewAllocation = 0;
PBCB Bcb = NULL;
BOOLEAN UnwindWeAllocatedDiskSpace = FALSE;
BOOLEAN UnwindAllocationSizeSet = FALSE;
BOOLEAN UnwindCacheManagerInformed = FALSE;
BOOLEAN UnwindWeInitializedMcb = FALSE;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatAddFileAllocation\n", 0);
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
DebugTrace( 0, Dbg, " DesiredAllocationSize = %8lx\n", DesiredAllocationSize);
Vcb = FcbOrDcb->Vcb;
//
// If we haven't yet set the correct AllocationSize, do so.
//
if (FcbOrDcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
FatLookupFileAllocationSize( IrpContext, FcbOrDcb );
}
//
// Check for the benign case that the desired allocation is already
// within the allocation size.
//
if (DesiredAllocationSize <= FcbOrDcb->Header.AllocationSize.LowPart) {
DebugTrace(0, Dbg, "Desired size within current allocation.\n", 0);
DebugTrace(-1, Dbg, "FatAddFileAllocation -> (VOID)\n", 0);
return;
}
DebugTrace( 0, Dbg, "InitialAllocation = %08lx.\n", FcbOrDcb->Header.AllocationSize.LowPart);
//
// Get a chunk of disk space that will fullfill our needs. If there
// was no initial allocation, start from the hint in the Vcb, otherwise
// try to allocate from the cluster after the initial allocation.
//
// If there was no initial allocation to the file, we can just use the
// Mcb in the FcbOrDcb, otherwise we have to use a new one, and merge
// it to the one in the FcbOrDcb.
//
_SEH2_TRY {
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
LBO FirstLboOfFile;
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
FatGetDirentFromFcbOrDcb( IrpContext,
FcbOrDcb,
FALSE,
&Dirent,
&Bcb );
//
// Set this dirty right now since this call can fail.
//
FatSetDirtyBcb( IrpContext, Bcb, Vcb, TRUE );
FatAllocateDiskSpace( IrpContext,
Vcb,
0,
&DesiredAllocationSize,
FALSE,
&FcbOrDcb->Mcb );
UnwindWeAllocatedDiskSpace = TRUE;
McbToCleanup = &FcbOrDcb->Mcb;
//
// We have to update the dirent and FcbOrDcb copies of
// FirstClusterOfFile since before it was 0
//
FatLookupMcbEntry( FcbOrDcb->Vcb,
&FcbOrDcb->Mcb,
0,
&FirstLboOfFile,
(PULONG)NULL,
NULL );
DebugTrace( 0, Dbg, "First Lbo of file will be %08lx.\n", FirstLboOfFile );
FcbOrDcb->FirstClusterOfFile = FatGetIndexFromLbo( Vcb, FirstLboOfFile );
Dirent->FirstClusterOfFile = (USHORT)FcbOrDcb->FirstClusterOfFile;
if ( FatIsFat32(Vcb) ) {
Dirent->FirstClusterOfFileHi = (USHORT)(FcbOrDcb->FirstClusterOfFile >> 16);
}
//
// Note the size of the allocation we need to tell the cache manager about.
//
NewAllocation = DesiredAllocationSize;
} else {
LBO LastAllocatedLbo;
VBO DontCare;
//
// Get the first cluster following the current allocation. It is possible
// the Mcb is empty (or short, etc.) so we need to be slightly careful
// about making sure we don't lie with the hint.
//
(void)FatLookupLastMcbEntry( FcbOrDcb->Vcb, &FcbOrDcb->Mcb, &DontCare, &LastAllocatedLbo, NULL );
//
// Try to get some disk space starting from there.
//
NewAllocation = DesiredAllocationSize - FcbOrDcb->Header.AllocationSize.LowPart;
FsRtlInitializeLargeMcb( &NewMcb, PagedPool );
UnwindWeInitializedMcb = TRUE;
McbToCleanup = &NewMcb;
FatAllocateDiskSpace( IrpContext,
Vcb,
(LastAllocatedLbo != ~0 ?
FatGetIndexFromLbo(Vcb,LastAllocatedLbo + 1) :
0),
&NewAllocation,
FALSE,
&NewMcb );
UnwindWeAllocatedDiskSpace = TRUE;
}
//
// Now that we increased the allocation of the file, mark it in the
// FcbOrDcb. Carefully prepare to handle an inability to grow the cache
// structures.
//
FcbOrDcb->Header.AllocationSize.LowPart += NewAllocation;
//
// Handle the maximal file case, where we may have just wrapped. Note
// that this must be the precise boundary case wrap, i.e. by one byte,
// so that the new allocation is actually one byte "less" as far as we're
// concerned. This is important for the extension case.
//
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
NewAllocation -= 1;
FcbOrDcb->Header.AllocationSize.LowPart = 0xffffffff;
}
UnwindAllocationSizeSet = TRUE;
//
// Inform the cache manager to increase the section size
//
if ( ARGUMENT_PRESENT(FileObject) && CcIsFileCached(FileObject) ) {
CcSetFileSizes( FileObject,
(PCC_FILE_SIZES)&FcbOrDcb->Header.AllocationSize );
UnwindCacheManagerInformed = TRUE;
}
//
// In the extension case, we have held off actually gluing the new
// allocation onto the file. This simplifies exception cleanup since
// if it was already added and the section grow failed, we'd have to
// do extra work to unglue it. This way, we can assume that if we
// raise the only thing we need to do is deallocate the disk space.
//
// Merge the allocation now.
//
if (FcbOrDcb->Header.AllocationSize.LowPart != NewAllocation) {
//
// Tack the new Mcb onto the end of the FcbOrDcb one.
//
FatMergeAllocation( IrpContext,
Vcb,
&FcbOrDcb->Mcb,
&NewMcb );
}
} _SEH2_FINALLY {
DebugUnwind( FatAddFileAllocation );
//
// Give FlushFileBuffer/Cleanup a clue here, regardless of success/fail..
//
SetFlag(FcbOrDcb->FcbState, FCB_STATE_FLUSH_FAT);
//
// If we were dogged trying to complete this operation, we need to go
// back various things out.
//
if (_SEH2_AbnormalTermination()) {
//
// Pull off the allocation size we tried to add to this object if
// we failed to grow cache structures or Mcb structures.
//
if (UnwindAllocationSizeSet) {
FcbOrDcb->Header.AllocationSize.LowPart -= NewAllocation;
}
if (UnwindCacheManagerInformed) {
CcSetFileSizes( FileObject,
(PCC_FILE_SIZES)&FcbOrDcb->Header.AllocationSize );
}
//
// In the case of initial allocation, we used the Fcb's Mcb and have
// to clean that up as well as the FAT chain references.
//
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
if (Dirent != NULL) {
FcbOrDcb->FirstClusterOfFile = 0;
Dirent->FirstClusterOfFile = 0;
if ( FatIsFat32(Vcb) ) {
Dirent->FirstClusterOfFileHi = 0;
}
}
}
//
// ... and drop the dirent Bcb if we got it. Do it now
// so we can afford to take the exception if we have to.
//
FatUnpinBcb( IrpContext, Bcb );
_SEH2_TRY {
//
// Note this can re-raise.
//
if ( UnwindWeAllocatedDiskSpace ) {
FatDeallocateDiskSpace( IrpContext, Vcb, McbToCleanup, FALSE );
}
} _SEH2_FINALLY {
//
// We always want to clean up the non-initial allocation temporary Mcb,
// otherwise we have the Fcb's Mcb and we just truncate it away.
//
if (UnwindWeInitializedMcb == TRUE) {
//
// Note that we already know a raise is in progress. No danger
// of encountering the normal case code below and doing this again.
//
FsRtlUninitializeLargeMcb( McbToCleanup );
} else {
if (McbToCleanup) {
FsRtlTruncateLargeMcb( McbToCleanup, 0 );
}
}
} _SEH2_END;
}
DebugTrace(-1, Dbg, "FatAddFileAllocation -> (VOID)\n", 0);
} _SEH2_END;
//
// Non-exceptional cleanup we always want to do. In handling the re-raise possibilities
// during exceptions we had to make sure these two steps always happened there beforehand.
// So now we handle the usual case.
//
FatUnpinBcb( IrpContext, Bcb );
if (UnwindWeInitializedMcb == TRUE) {
FsRtlUninitializeLargeMcb( &NewMcb );
}
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatTruncateFileAllocation (
IN PIRP_CONTEXT IrpContext,
IN PFCB FcbOrDcb,
IN ULONG DesiredAllocationSize
)
/*++
Routine Description:
This routine truncates the allocation to the specified file/directory.
If the file is already smaller than the indicated size then this procedure
is effectively a noop.
Arguments:
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified
This parameter must not specify the root dcb.
DesiredAllocationSize - Supplies the maximum size, in bytes, that we want
allocated to the file/directory. It is rounded
up to the nearest cluster.
Return Value:
VOID - TRUE if the operation completed and FALSE if it had to
block but could not.
--*/
{
PVCB Vcb;
PBCB Bcb = NULL;
LARGE_MCB RemainingMcb = {0};
ULONG BytesPerCluster;
PDIRENT Dirent = NULL;
BOOLEAN UpdatedDirent = FALSE;
ULONG UnwindInitialAllocationSize;
ULONG UnwindInitialFirstClusterOfFile;
BOOLEAN UnwindWeAllocatedMcb = FALSE;
PAGED_CODE();
Vcb = FcbOrDcb->Vcb;
DebugTrace(+1, Dbg, "FatTruncateFileAllocation\n", 0);
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
DebugTrace( 0, Dbg, " DesiredAllocationSize = %8lx\n", DesiredAllocationSize);
//
// If the Fcb isn't in good condition, we have no business whacking around on
// the disk after "its" clusters.
//
// Inspired by a Prefix complaint.
//
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
//
// If we haven't yet set the correct AllocationSize, do so.
//
if (FcbOrDcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
FatLookupFileAllocationSize( IrpContext, FcbOrDcb );
}
//
// Round up the Desired Allocation Size to the next cluster size
//
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
//
// Note if the desired allocation is zero, to distinguish this from
// the wrap case below.
//
if (DesiredAllocationSize != 0) {
DesiredAllocationSize = (DesiredAllocationSize + (BytesPerCluster - 1)) &
~(BytesPerCluster - 1);
//
// Check for the benign case that the file is already smaller than
// the desired truncation. Note that if it wraps, then a) it was
// specifying an offset in the maximally allocatable cluster and
// b) we're not asking to extend the file, either. So stop.
//
if (DesiredAllocationSize == 0 ||
DesiredAllocationSize >= FcbOrDcb->Header.AllocationSize.LowPart) {
DebugTrace(0, Dbg, "Desired size within current allocation.\n", 0);
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
return;
}
}
//
// This is a no-op if the allocation size is already what we want.
//
if (DesiredAllocationSize == FcbOrDcb->Header.AllocationSize.LowPart) {
DebugTrace(0, Dbg, "Desired size equals current allocation.\n", 0);
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
return;
}
UnwindInitialAllocationSize = FcbOrDcb->Header.AllocationSize.LowPart;
UnwindInitialFirstClusterOfFile = FcbOrDcb->FirstClusterOfFile;
//
// Update the FcbOrDcb allocation size. If it is now zero, we have the
// additional task of modifying the FcbOrDcb and Dirent copies of
// FirstClusterInFile.
//
// Note that we must pin the dirent before actually deallocating the
// disk space since, in unwind, it would not be possible to reallocate
// deallocated disk space as someone else may have reallocated it and
// may cause an exception when you try to get some more disk space.
// Thus FatDeallocateDiskSpace must be the final dangerous operation.
//
_SEH2_TRY {
FcbOrDcb->Header.AllocationSize.QuadPart = DesiredAllocationSize;
//
// Special case 0
//
if (DesiredAllocationSize == 0) {
//
// We have to update the dirent and FcbOrDcb copies of
// FirstClusterOfFile since before it was 0
//
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
FatGetDirentFromFcbOrDcb( IrpContext, FcbOrDcb, FALSE, &Dirent, &Bcb );
Dirent->FirstClusterOfFile = 0;
if (FatIsFat32(Vcb)) {
Dirent->FirstClusterOfFileHi = 0;
}
FcbOrDcb->FirstClusterOfFile = 0;
FatSetDirtyBcb( IrpContext, Bcb, Vcb, TRUE );
UpdatedDirent = TRUE;
FatDeallocateDiskSpace( IrpContext, Vcb, &FcbOrDcb->Mcb, ((FcbOrDcb->FcbState & FCB_STATE_ZERO_ON_DEALLOCATION) != 0));
FatRemoveMcbEntry( FcbOrDcb->Vcb, &FcbOrDcb->Mcb, 0, 0xFFFFFFFF );
} else {
//
// Split the existing allocation into two parts, one we will keep, and
// one we will deallocate.
//
FsRtlInitializeLargeMcb( &RemainingMcb, PagedPool );
UnwindWeAllocatedMcb = TRUE;
FatSplitAllocation( IrpContext,
Vcb,
&FcbOrDcb->Mcb,
DesiredAllocationSize,
&RemainingMcb );
FatDeallocateDiskSpace( IrpContext, Vcb, &RemainingMcb, ((FcbOrDcb->FcbState & FCB_STATE_ZERO_ON_DEALLOCATION) != 0) );
FsRtlUninitializeLargeMcb( &RemainingMcb );
}
} _SEH2_FINALLY {
DebugUnwind( FatTruncateFileAllocation );
//
// Is this really the right backout strategy? It would be nice if we could
// pretend the truncate worked if we knew that the file had gotten into
// a consistent state. Leaving dangled clusters is probably quite preferable.
//
if ( _SEH2_AbnormalTermination() ) {
FcbOrDcb->Header.AllocationSize.LowPart = UnwindInitialAllocationSize;
if ( (DesiredAllocationSize == 0) && (Dirent != NULL)) {
if (UpdatedDirent) {
//
// If the dirent has been updated ok and marked dirty, then we
// failed in deallocatediscspace, and don't know what state
// the on disc fat chain is in. So we throw away the mcb,
// and potentially loose a few clusters until the next
// chkdsk. The operation has succeeded, but the exception
// will still propogate. 5.1
//
FatRemoveMcbEntry( Vcb, &FcbOrDcb->Mcb, 0, 0xFFFFFFFF );
FcbOrDcb->Header.AllocationSize.QuadPart = 0;
}
else if (FcbOrDcb->FirstClusterOfFile == 0) {
Dirent->FirstClusterOfFile = (USHORT)UnwindInitialFirstClusterOfFile;
if ( FatIsFat32(Vcb) ) {
Dirent->FirstClusterOfFileHi =
(USHORT)(UnwindInitialFirstClusterOfFile >> 16);
}
FcbOrDcb->FirstClusterOfFile = UnwindInitialFirstClusterOfFile;
}
}
if ( UnwindWeAllocatedMcb ) {
FsRtlUninitializeLargeMcb( &RemainingMcb );
}
//
// Note that in the non zero truncation case, we will also
// leak clusters. However, apart from this, the in memory and on disc
// structures will agree.
}
FatUnpinBcb( IrpContext, Bcb );
//
// Give FlushFileBuffer/Cleanup a clue here, regardless of success/fail.
//
SetFlag(FcbOrDcb->FcbState, FCB_STATE_FLUSH_FAT);
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
} _SEH2_END;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatLookupFileAllocationSize (
IN PIRP_CONTEXT IrpContext,
IN PFCB FcbOrDcb
)
/*++
Routine Description:
This routine retrieves the current file allocatio size for the
specified file/directory.
Arguments:
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified
--*/
{
LBO Lbo;
ULONG ByteCount;
BOOLEAN DontCare;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatLookupAllocationSize\n", 0);
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
//
// We call FatLookupFileAllocation with Vbo of 0xffffffff - 1.
//
FatLookupFileAllocation( IrpContext,
FcbOrDcb,
MAXULONG - 1,
&Lbo,
&ByteCount,
&DontCare,
&DontCare,
NULL );
//
// FileSize was set at Fcb creation time from the contents of the directory entry,
// and we are only now looking up the real length of the allocation chain. If it
// cannot be contained, this is trash. Probably more where that came from.
//
if (FcbOrDcb->Header.FileSize.LowPart > FcbOrDcb->Header.AllocationSize.LowPart) {
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
}
DebugTrace(-1, Dbg, "FatLookupFileAllocationSize -> (VOID)\n", 0);
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatAllocateDiskSpace (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG AbsoluteClusterHint,
IN PULONG ByteCount,
IN BOOLEAN ExactMatchRequired,
OUT PLARGE_MCB Mcb
)
/*++
Routine Description:
This procedure allocates additional disk space and builds an mcb
representing the newly allocated space. If the space cannot be
allocated then this procedure raises an appropriate status.
Searching starts from the hint index in the Vcb unless an alternative
non-zero hint is given in AlternateClusterHint. If we are using the
hint field in the Vcb, it is set to the cluster following our allocation
when we are done.
Disk space can only be allocated in cluster units so this procedure
will round up any byte count to the next cluster boundary.
Pictorially what is done is the following (where ! denotes the end of
the fat chain (i.e., FAT_CLUSTER_LAST)):
Mcb (empty)
becomes
Mcb |--a--|--b--|--c--!
^
ByteCount ----------+
Arguments:
Vcb - Supplies the VCB being modified
AbsoluteClusterHint - Supplies an alternate hint index to start the
search from. If this is zero we use, and update,
the Vcb hint field.
ByteCount - Supplies the number of bytes that we are requesting, and
receives the number of bytes that we got.
ExactMatchRequired - Caller should set this to TRUE if only the precise run requested
is acceptable.
Mcb - Receives the MCB describing the newly allocated disk space. The
caller passes in an initialized Mcb that is filled in by this procedure.
Return Value:
TRUE - Allocated ok
FALSE - Failed to allocate exactly as requested (=> ExactMatchRequired was TRUE)
--*/
{
UCHAR LogOfBytesPerCluster;
ULONG BytesPerCluster;
ULONG StartingCluster;
ULONG ClusterCount;
ULONG WindowRelativeHint;
#if DBG
ULONG PreviousClear = 0;
#endif
PFAT_WINDOW Window;
BOOLEAN Wait = FALSE;
BOOLEAN Result = TRUE;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatAllocateDiskSpace\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " *ByteCount = %8lx\n", *ByteCount);
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
DebugTrace( 0, Dbg, " Hint = %8lx\n", AbsoluteClusterHint);
NT_ASSERT((AbsoluteClusterHint <= Vcb->AllocationSupport.NumberOfClusters + 2) && (1 != AbsoluteClusterHint));
//
// Make sure byte count is not zero
//
if (*ByteCount == 0) {
DebugTrace(0, Dbg, "Nothing to allocate.\n", 0);
DebugTrace(-1, Dbg, "FatAllocateDiskSpace -> (VOID)\n", 0);
return;
}
//
// Compute the cluster count based on the byte count, rounding up
// to the next cluster if there is any remainder. Note that the
// pathalogical case BytesCount == 0 has been eliminated above.
//
LogOfBytesPerCluster = Vcb->AllocationSupport.LogOfBytesPerCluster;
BytesPerCluster = 1 << LogOfBytesPerCluster;
*ByteCount = (*ByteCount + (BytesPerCluster - 1))
& ~(BytesPerCluster - 1);
//
// If ByteCount is NOW zero, then we were asked for the maximal
// filesize (or at least for bytes in the last allocatable sector).
//
if (*ByteCount == 0) {
*ByteCount = 0xffffffff;
ClusterCount = 1 << (32 - LogOfBytesPerCluster);
} else {
ClusterCount = (*ByteCount >> LogOfBytesPerCluster);
}
//
// Analysis tools don't figure out that ClusterCount is not zero because
// of the ByteCount == 0 checks, so give them a hint.
//
_Analysis_assume_(ClusterCount > 0);
//
// Make sure there are enough free clusters to start with, and
// take them now so that nobody else takes them from us.
//
ExAcquireResourceSharedLite(&Vcb->ChangeBitMapResource, TRUE);
FatLockFreeClusterBitMap( Vcb );
if (ClusterCount <= Vcb->AllocationSupport.NumberOfFreeClusters) {
Vcb->AllocationSupport.NumberOfFreeClusters -= ClusterCount;
} else {
FatUnlockFreeClusterBitMap( Vcb );
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
DebugTrace(0, Dbg, "Disk Full. Raise Status.\n", 0);
FatRaiseStatus( IrpContext, STATUS_DISK_FULL );
}
//
// Did the caller supply a hint?
//
if ((0 != AbsoluteClusterHint) && (AbsoluteClusterHint < (Vcb->AllocationSupport.NumberOfClusters + 2))) {
if (Vcb->NumberOfWindows > 1) {
//
// If we're being called upon to allocate clusters outside the
// current window (which happens only via MoveFile), it's a problem.
// We address this by changing the current window to be the one which
// contains the alternate cluster hint. Note that if the user's
// request would cross a window boundary, he doesn't really get what
// he wanted.
//
if (AbsoluteClusterHint < Vcb->CurrentWindow->FirstCluster ||
AbsoluteClusterHint > Vcb->CurrentWindow->LastCluster) {
ULONG BucketNum = FatWindowOfCluster( AbsoluteClusterHint );
NT_ASSERT( BucketNum < Vcb->NumberOfWindows);
//
// Drop our shared lock on the ChangeBitMapResource, and pick it up again
// exclusive in preparation for making the window swap.
//
FatUnlockFreeClusterBitMap(Vcb);
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
ExAcquireResourceExclusiveLite(&Vcb->ChangeBitMapResource, TRUE);
FatLockFreeClusterBitMap(Vcb);
Window = &Vcb->Windows[BucketNum];
//
// Again, test the current window against the one we want - some other
// thread could have sneaked in behind our backs and kindly set it to the one
// we need, when we dropped and reacquired the ChangeBitMapResource above.
//
if (Window != Vcb->CurrentWindow) {
_SEH2_TRY {
Wait = BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
SetFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
//
// Change to the new window (update Vcb->CurrentWindow) and scan it
// to build up a freespace bitmap etc.
//
FatExamineFatEntries( IrpContext, Vcb,
0,
0,
FALSE,
Window,
NULL);
} _SEH2_FINALLY {
if (!Wait) {
ClearFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
}
if (_SEH2_AbnormalTermination()) {
//
// We will have raised as a result of failing to pick up the
// chunk of the FAT for this window move. Release our resources
// and return the cluster count to the volume.
//
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
FatUnlockFreeClusterBitMap( Vcb );
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
}
} _SEH2_END;
}
}
//
// Make the hint cluster number relative to the base of the current window...
//
// Currentwindow->Firstcluster is baised by +2 already, so we will lose the
// bias already in AbsoluteClusterHint. Put it back....
//
WindowRelativeHint = AbsoluteClusterHint - Vcb->CurrentWindow->FirstCluster + 2;
}
else {
//
// Only one 'window', ie fat16/12. No modification necessary.
//
WindowRelativeHint = AbsoluteClusterHint;
}
}
else {
//
// Either no hint supplied, or it was out of range, so grab one from the Vcb
//
// NOTE: Clusterhint in the Vcb is not guaranteed to be set (may be -1)
//
WindowRelativeHint = Vcb->ClusterHint;
AbsoluteClusterHint = 0;
//
// Vcb hint may not have been initialized yet. Force to valid cluster.
//
if (-1 == WindowRelativeHint) {
WindowRelativeHint = 2;
}
}
NT_ASSERT((WindowRelativeHint >= 2) && (WindowRelativeHint < Vcb->FreeClusterBitMap.SizeOfBitMap + 2));
//
// Keep track of the window we're allocating from, so we can clean
// up correctly if the current window changes after we unlock the
// bitmap.
//
Window = Vcb->CurrentWindow;
//
// Try to find a run of free clusters large enough for us.
//
StartingCluster = FatFindFreeClusterRun( IrpContext,
Vcb,
ClusterCount,
WindowRelativeHint );
//
// If the above call was successful, we can just update the fat
// and Mcb and exit. Otherwise we have to look for smaller free
// runs.
//
// This test is a bit funky. Note that the error return from
// RtlFindClearBits is -1, and adding two to that is 1.
//
if ((StartingCluster != 1) &&
((0 == AbsoluteClusterHint) || (StartingCluster == WindowRelativeHint))
) {
#if DBG
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
#endif // DBG
//
// Take the clusters we found, and unlock the bit map.
//
FatReserveClusters(IrpContext, Vcb, StartingCluster, ClusterCount);
Window->ClustersFree -= ClusterCount;
StartingCluster += Window->FirstCluster;
StartingCluster -= 2;
NT_ASSERT( PreviousClear - ClusterCount == Window->ClustersFree );
FatUnlockFreeClusterBitMap( Vcb );
//
// Note that this call will never fail since there is always
// room for one entry in an empty Mcb.
//
FatAddMcbEntry( Vcb, Mcb,
0,
FatGetLboFromIndex( Vcb, StartingCluster ),
*ByteCount);
_SEH2_TRY {
//
// Update the fat.
//
FatAllocateClusters(IrpContext, Vcb,
StartingCluster,
ClusterCount);
} _SEH2_FINALLY {
DebugUnwind( FatAllocateDiskSpace );
//
// If the allocate clusters failed, remove the run from the Mcb,
// unreserve the clusters, and reset the free cluster count.
//
if (_SEH2_AbnormalTermination()) {
FatRemoveMcbEntry( Vcb, Mcb, 0, *ByteCount );
FatLockFreeClusterBitMap( Vcb );
// Only clear bits if the bitmap window is the same.
if (Window == Vcb->CurrentWindow) {
// Both values (startingcluster and window->firstcluster) are
// already biased by 2, so will cancel, so we need to add in the 2 again.
FatUnreserveClusters( IrpContext, Vcb,
StartingCluster - Window->FirstCluster + 2,
ClusterCount );
}
Window->ClustersFree += ClusterCount;
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
FatUnlockFreeClusterBitMap( Vcb );
}
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
} _SEH2_END;
} else {
//
// Note that Index is a zero-based window-relative number. When appropriate
// it'll get converted into a true cluster number and put in Cluster, which
// will be a volume relative true cluster number.
//
ULONG Index = 0;
ULONG Cluster = 0;
ULONG CurrentVbo = 0;
ULONG PriorLastCluster = 0;
ULONG BytesFound = 0;
ULONG ClustersFound = 0;
ULONG ClustersRemaining = 0;
BOOLEAN LockedBitMap = FALSE;
BOOLEAN SelectNextContigWindow = FALSE;
//
// Drop our shared lock on the ChangeBitMapResource, and pick it up again
// exclusive in preparation for making a window swap.
//
FatUnlockFreeClusterBitMap(Vcb);
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
ExAcquireResourceExclusiveLite(&Vcb->ChangeBitMapResource, TRUE);
FatLockFreeClusterBitMap(Vcb);
LockedBitMap = TRUE;
_SEH2_TRY {
if ( ExactMatchRequired && (1 == Vcb->NumberOfWindows)) {
//
// Give up right now, there are no more windows to search! RtlFindClearBits
// searchs the whole bitmap, so we would have found any contiguous run
// large enough.
//
try_leave( Result = FALSE);
}
//
// While the request is still incomplete, look for the largest
// run of free clusters, mark them taken, allocate the run in
// the Mcb and Fat, and if this isn't the first time through
// the loop link it to prior run on the fat. The Mcb will
// coalesce automatically.
//
ClustersRemaining = ClusterCount;
CurrentVbo = 0;
PriorLastCluster = 0;
while (ClustersRemaining != 0) {
//
// If we just entered the loop, the bit map is already locked
//
if ( !LockedBitMap ) {
FatLockFreeClusterBitMap( Vcb );
LockedBitMap = TRUE;
}
//
// Find the largest run of free clusters. If the run is
// bigger than we need, only use what we need. Note that
// this will then be the last while() iteration.
//
// 12/3/95: need to bias bitmap by 2 bits for the defrag
// hooks and the below macro became impossible to do without in-line
// procedures.
//
// ClustersFound = FatLongestFreeClusterRun( IrpContext, Vcb, &Index );
ClustersFound = 0;
if (!SelectNextContigWindow) {
if ( 0 != WindowRelativeHint) {
ULONG Desired = Vcb->FreeClusterBitMap.SizeOfBitMap - (WindowRelativeHint - 2);
//
// We will try to allocate contiguously. Try from the current hint the to
// end of current window. Don't try for more than we actually need.
//
if (Desired > ClustersRemaining) {
Desired = ClustersRemaining;
}
if (RtlAreBitsClear( &Vcb->FreeClusterBitMap,
WindowRelativeHint - 2,
Desired))
{
//
// Clusters from hint->...windowend are free. Take them.
//
Index = WindowRelativeHint - 2;
ClustersFound = Desired;
if (FatIsFat32(Vcb)) {
//
// We're now up against the end of the current window, so indicate that we
// want the next window in the sequence next time around. (If we're not up
// against the end of the window, then we got what we needed and won't be
// coming around again anyway).
//
SelectNextContigWindow = TRUE;
WindowRelativeHint = 2;
}
else {
//
// FAT 12/16 - we've run up against the end of the volume. Clear the
// hint, since we now have no idea where to look.
//
WindowRelativeHint = 0;
}
#if DBG
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
#endif // DBG
}
else {
if (ExactMatchRequired) {
//
// If our caller required an exact match, then we're hosed. Bail out now.
//
try_leave( Result = FALSE);
}
//
// Hint failed, drop back to pot luck
//
WindowRelativeHint = 0;
}
}
if ((0 == WindowRelativeHint) && (0 == ClustersFound)) {
if (ClustersRemaining <= Vcb->CurrentWindow->ClustersFree) {
//
// The remaining allocation could be satisfied entirely from this
// window. We will ask only for what we need, to try and avoid
// unnecessarily fragmenting large runs of space by always using
// (part of) the largest run we can find. This call will return the
// first run large enough.
//
Index = RtlFindClearBits( &Vcb->FreeClusterBitMap, ClustersRemaining, 0);
if (-1 != Index) {
ClustersFound = ClustersRemaining;
}
}
if (0 == ClustersFound) {
//
// Still nothing, so just take the largest free run we can find.
//
ClustersFound = RtlFindLongestRunClear( &Vcb->FreeClusterBitMap, &Index );
}
#if DBG
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
#endif // DBG
if (ClustersFound >= ClustersRemaining) {
ClustersFound = ClustersRemaining;
}
else {
//
// If we just ran up to the end of a window, set up a hint that
// we'd like the next consecutive window after this one. (FAT32 only)
//
if ( ((Index + ClustersFound) == Vcb->FreeClusterBitMap.SizeOfBitMap) &&
FatIsFat32( Vcb)
) {
SelectNextContigWindow = TRUE;
WindowRelativeHint = 2;
}
}
}
}
if (ClustersFound == 0) {
ULONG FaveWindow = 0;
BOOLEAN SelectedWindow;
//
// If we found no free clusters on a single-window FAT,
// there was a bad problem with the free cluster count.
//
if (1 == Vcb->NumberOfWindows) {
#ifdef _MSC_VER
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
#endif
FatBugCheck( 0, 5, 0 );
}
//
// Switch to a new bucket. Possibly the next one if we're
// currently on a roll (allocating contiguously)
//
SelectedWindow = FALSE;
if ( SelectNextContigWindow) {
ULONG NextWindow;
NextWindow = (((ULONG)((PUCHAR)Vcb->CurrentWindow - (PUCHAR)Vcb->Windows)) / sizeof( FAT_WINDOW)) + 1;
if ((NextWindow < Vcb->NumberOfWindows) &&
( Vcb->Windows[ NextWindow].ClustersFree > 0)
) {
FaveWindow = NextWindow;
SelectedWindow = TRUE;
}
else {
if (ExactMatchRequired) {
//
// Some dope tried to allocate a run past the end of the volume...
//
try_leave( Result = FALSE);
}
//
// Give up on the contiguous allocation attempts
//
WindowRelativeHint = 0;
}
SelectNextContigWindow = FALSE;
}
if (!SelectedWindow) {
//
// Select a new window to begin allocating from
//
FaveWindow = FatSelectBestWindow( Vcb);
}
//
// By now we'd better have found a window with some free clusters
//
if (0 == Vcb->Windows[ FaveWindow].ClustersFree) {
#ifdef _MSC_VER
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
#endif
FatBugCheck( 0, 5, 1 );
}
Wait = BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
SetFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
FatExamineFatEntries( IrpContext, Vcb,
0,
0,
FALSE,
&Vcb->Windows[FaveWindow],
NULL);
if (!Wait) {
ClearFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
}
//
// Now we'll just go around the loop again, having switched windows,
// and allocate....
//
#if DBG
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
#endif //DBG
} // if (clustersfound == 0)
else {
//
// Take the clusters we found, convert our index to a cluster number
// and unlock the bit map.
//
Window = Vcb->CurrentWindow;
FatReserveClusters( IrpContext, Vcb, (Index + 2), ClustersFound );
Cluster = Index + Window->FirstCluster;
Window->ClustersFree -= ClustersFound;
NT_ASSERT( PreviousClear - ClustersFound == Window->ClustersFree );
FatUnlockFreeClusterBitMap( Vcb );
LockedBitMap = FALSE;
//
// Add the newly alloced run to the Mcb.
//
BytesFound = ClustersFound << LogOfBytesPerCluster;
FatAddMcbEntry( Vcb, Mcb,
CurrentVbo,
FatGetLboFromIndex( Vcb, Cluster ),
BytesFound );
//
// Connect the last allocated run with this one, and allocate
// this run on the Fat.
//
if (PriorLastCluster != 0) {
FatSetFatEntry( IrpContext,
Vcb,
PriorLastCluster,
(FAT_ENTRY)Cluster );
}
//
// Update the fat
//
FatAllocateClusters( IrpContext, Vcb, Cluster, ClustersFound );
//
// Prepare for the next iteration.
//
CurrentVbo += BytesFound;
ClustersRemaining -= ClustersFound;
PriorLastCluster = Cluster + ClustersFound - 1;
}
} // while (clustersremaining)
} _SEH2_FINALLY {
DebugUnwind( FatAllocateDiskSpace );
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
//
// Is there any unwinding to do?
//
if ( _SEH2_AbnormalTermination() || (FALSE == Result)) {
//
// Flag to the caller that they're getting nothing
//
*ByteCount = 0;
//
// There are three places we could have taken this exception:
// when switching the window (FatExamineFatEntries), adding
// a found run to the Mcb (FatAddMcbEntry), or when writing
// the changes to the FAT (FatSetFatEntry). In the first case
// we don't have anything to unwind before deallocation, and
// can detect this by seeing if we have the ClusterBitmap
// mutex out.
if (!LockedBitMap) {
FatLockFreeClusterBitMap( Vcb );
//
// In these cases, we have the possiblity that the FAT
// window is still in place and we need to clear the bits.
// If the Mcb entry isn't there (we raised trying to add
// it), the effect of trying to remove it is a noop.
//
if (Window == Vcb->CurrentWindow) {
//
// Cluster reservation works on cluster 2 based window-relative
// numbers, so we must convert. The subtraction will lose the
// cluster 2 base, so bias the result.
//
FatUnreserveClusters( IrpContext, Vcb,
(Cluster - Window->FirstCluster) + 2,
ClustersFound );
}
//
// Note that FatDeallocateDiskSpace will take care of adjusting
// to account for the entries in the Mcb. All we have to account
// for is the last run that didn't make it.
//
Window->ClustersFree += ClustersFound;
Vcb->AllocationSupport.NumberOfFreeClusters += ClustersFound;
FatUnlockFreeClusterBitMap( Vcb );
FatRemoveMcbEntry( Vcb, Mcb, CurrentVbo, BytesFound );
} else {
//
// Just drop the mutex now - we didn't manage to do anything
// that needs to be backed out.
//
FatUnlockFreeClusterBitMap( Vcb );
}
_SEH2_TRY {
//
// Now we have tidied up, we are ready to just send the Mcb
// off to deallocate disk space
//
FatDeallocateDiskSpace( IrpContext, Vcb, Mcb, FALSE );
} _SEH2_FINALLY {
//
// Now finally (really), remove all the entries from the mcb
//
FatRemoveMcbEntry( Vcb, Mcb, 0, 0xFFFFFFFF );
} _SEH2_END;
}
DebugTrace(-1, Dbg, "FatAllocateDiskSpace -> (VOID)\n", 0);
} _SEH2_END; // finally
}
return;
}
//
// Limit our zeroing writes to 1 MB.
//
#define MAX_ZERO_MDL_SIZE (1*1024*1024)
_Requires_lock_held_(_Global_critical_region_)
VOID
FatDeallocateDiskSpace (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN PLARGE_MCB Mcb,
IN BOOLEAN ZeroOnDeallocate
)
/*++
Routine Description:
This procedure deallocates the disk space denoted by an input
mcb. Note that the input MCB does not need to necessarily describe
a chain that ends with a FAT_CLUSTER_LAST entry.
Pictorially what is done is the following
Fat |--a--|--b--|--c--|
Mcb |--a--|--b--|--c--|
becomes
Fat |--0--|--0--|--0--|
Mcb |--a--|--b--|--c--|
Arguments:
Vcb - Supplies the VCB being modified
Mcb - Supplies the MCB describing the disk space to deallocate. Note
that Mcb is unchanged by this procedure.
Return Value:
None.
--*/
{
LBO Lbo;
VBO Vbo;
ULONG RunsInMcb;
ULONG ByteCount;
ULONG ClusterCount = 0;
ULONG ClusterIndex = 0;
ULONG McbIndex = 0;
UCHAR LogOfBytesPerCluster;
PFAT_WINDOW Window;
NTSTATUS ZeroingStatus = STATUS_SUCCESS;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatDeallocateDiskSpace\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
LogOfBytesPerCluster = Vcb->AllocationSupport.LogOfBytesPerCluster;
RunsInMcb = FsRtlNumberOfRunsInLargeMcb( Mcb );
if ( RunsInMcb == 0 ) {
DebugTrace(-1, Dbg, "FatDeallocateDiskSpace -> (VOID)\n", 0);
return;
}
//
// If we are supposed to zero out the allocation before freeing it, do so.
//
if (ZeroOnDeallocate) {
_SEH2_TRY {
PIRP IoIrp;
KEVENT IoEvent;
IO_STATUS_BLOCK Iosb;
PVOID Buffer = NULL;
PMDL Mdl;
ULONG ByteCountToZero;
ULONG MdlSizeMapped;
//
// Issue the writes down for each run in the Mcb
//
KeInitializeEvent( &IoEvent,
NotificationEvent,
FALSE );
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
//
// Assert that Fat files have no holes.
//
NT_ASSERT( Lbo != 0 );
//
// Setup our MDL for the this run.
//
if (ByteCount > MAX_ZERO_MDL_SIZE) {
Mdl = FatBuildZeroMdl( IrpContext, MAX_ZERO_MDL_SIZE);
} else {
Mdl = FatBuildZeroMdl( IrpContext, ByteCount);
}
if (!Mdl) {
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
goto try_exit;
}
_SEH2_TRY {
//
// Map the MDL.
//
Buffer = MmGetSystemAddressForMdlSafe(Mdl, HighPagePriority|MdlMappingNoExecute);
if (!Buffer) {
NT_ASSERT( FALSE );
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
goto try_exit2;
}
//
// We might not have not been able to get an MDL big enough to map the whole
// run. In this case, break up the write.
//
MdlSizeMapped = min( ByteCount, Mdl->ByteCount );
ByteCountToZero = ByteCount;
//
// Loop until there are no bytes left to write
//
while (ByteCountToZero != 0) {
//
// Write zeros to each run.
//
KeClearEvent( &IoEvent );
IoIrp = IoBuildSynchronousFsdRequest( IRP_MJ_WRITE,
Vcb->TargetDeviceObject,
Buffer,
MdlSizeMapped,
(PLARGE_INTEGER)&Lbo,
&IoEvent,
&Iosb );
if (IoIrp == NULL) {
NT_ASSERT( FALSE );
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
goto try_exit2;
}
//
// Set a flag indicating that we want to write through any
// cache on the controller. This eliminates the need for
// an explicit flush-device after the write.
//
SetFlag( IoGetNextIrpStackLocation(IoIrp)->Flags, SL_WRITE_THROUGH );
ZeroingStatus = IoCallDriver( Vcb->TargetDeviceObject, IoIrp );
if (ZeroingStatus == STATUS_PENDING) {
(VOID)KeWaitForSingleObject( &IoEvent,
Executive,
KernelMode,
FALSE,
(PLARGE_INTEGER)NULL );
ZeroingStatus = Iosb.Status;
}
if (!NT_SUCCESS( ZeroingStatus )) {
NT_ASSERT( FALSE );
goto try_exit2;
}
//
// Increment the starting offset where we will zero.
//
Lbo += MdlSizeMapped;
//
// Decrement ByteCount
//
ByteCountToZero -= MdlSizeMapped;
if (ByteCountToZero < MdlSizeMapped) {
MdlSizeMapped = ByteCountToZero;
}
}
try_exit2:
NOTHING;
} _SEH2_FINALLY {
if (!FlagOn( Mdl->MdlFlags, MDL_SOURCE_IS_NONPAGED_POOL) &&
FlagOn( Mdl->MdlFlags, MDL_MAPPED_TO_SYSTEM_VA )) {
MmUnmapLockedPages( Mdl->MappedSystemVa, Mdl );
}
IoFreeMdl( Mdl );
} _SEH2_END;
}
try_exit:
NOTHING;
} _SEH2_EXCEPT(FatExceptionFilter( NULL, _SEH2_GetExceptionInformation() )) {
//
// If we failed to zero for some reason, still go ahead and deallocate
// the clusters. Otherwise we'll leak space from the volume.
//
ZeroingStatus = _SEH2_GetExceptionCode();
} _SEH2_END;
}
NT_ASSERT( NT_SUCCESS(ZeroingStatus) );
_SEH2_TRY {
//
// Run though the Mcb, freeing all the runs in the fat.
//
// We do this in two steps (first update the fat, then the bitmap
// (which can't fail)) to prevent other people from taking clusters
// that we need to re-allocate in the event of unwind.
//
ExAcquireResourceSharedLite(&Vcb->ChangeBitMapResource, TRUE);
RunsInMcb = FsRtlNumberOfRunsInLargeMcb( Mcb );
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
//
// Assert that Fat files have no holes.
//
NT_ASSERT( Lbo != 0 );
//
// Write FAT_CLUSTER_AVAILABLE to each cluster in the run.
//
if (ByteCount == 0xFFFFFFFF) {
//
// Special case the computation of ClusterCout
// when file is of max size (4GiB - 1).
//
ClusterCount = (1 << (32 - LogOfBytesPerCluster));
} else {
ClusterCount = ByteCount >> LogOfBytesPerCluster;
}
ClusterIndex = FatGetIndexFromLbo( Vcb, Lbo );
FatFreeClusters( IrpContext, Vcb, ClusterIndex, ClusterCount );
}
//
// From now on, nothing can go wrong .... (as in raise)
//
FatLockFreeClusterBitMap( Vcb );
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
ULONG ClusterEnd;
ULONG MyStart, MyLength, count;
#if DBG
#ifndef __REACTOS__
ULONG PreviousClear = 0;
#endif
ULONG i = 0;
#endif
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
//
// Mark the bits clear in the FreeClusterBitMap.
//
if (ByteCount == 0xFFFFFFFF) {
//
// Special case the computation of ClusterCout
// when file is of max size (2^32 - 1).
//
ClusterCount = (1 << (32 - LogOfBytesPerCluster));
} else {
ClusterCount = ByteCount >> LogOfBytesPerCluster;
}
ClusterIndex = FatGetIndexFromLbo( Vcb, Lbo );
Window = Vcb->CurrentWindow;
//
// If we've divided the bitmap, elide bitmap manipulation for
// runs that are outside the current bucket.
//
ClusterEnd = ClusterIndex + ClusterCount - 1;
if (!(ClusterIndex > Window->LastCluster ||
ClusterEnd < Window->FirstCluster)) {
//
// The run being freed overlaps the current bucket, so we'll
// have to clear some bits.
//
if (ClusterIndex < Window->FirstCluster &&
ClusterEnd > Window->LastCluster) {
MyStart = Window->FirstCluster;
MyLength = Window->LastCluster - Window->FirstCluster + 1;
} else if (ClusterIndex < Window->FirstCluster) {
MyStart = Window->FirstCluster;
MyLength = ClusterEnd - Window->FirstCluster + 1;
} else {
//
// The range being freed starts in the bucket, and may possibly
// extend beyond the bucket.
//
MyStart = ClusterIndex;
if (ClusterEnd <= Window->LastCluster) {
MyLength = ClusterCount;
} else {
MyLength = Window->LastCluster - ClusterIndex + 1;
}
}
if (MyLength == 0) {
continue;
}
#if DBG
#ifndef __REACTOS__
#ifdef _MSC_VER
#pragma prefast( suppress:28931, "this is DBG build only" )
#endif
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
#endif
//
// Verify that the Bits are all really set.
//
NT_ASSERT( MyStart + MyLength - Window->FirstCluster <= Vcb->FreeClusterBitMap.SizeOfBitMap );
for (i = 0; i < MyLength; i++) {
NT_ASSERT( RtlCheckBit(&Vcb->FreeClusterBitMap,
MyStart - Window->FirstCluster + i) == 1 );
}
#endif // DBG
FatUnreserveClusters( IrpContext, Vcb,
MyStart - Window->FirstCluster + 2,
MyLength );
}
//
// Adjust the ClustersFree count for each bitmap window, even the ones
// that are not the current window.
//
if (FatIsFat32(Vcb)) {
Window = &Vcb->Windows[FatWindowOfCluster( ClusterIndex )];
} else {
Window = &Vcb->Windows[0];
}
MyStart = ClusterIndex;
for (MyLength = ClusterCount; MyLength > 0; MyLength -= count) {
count = FatMin(Window->LastCluster - MyStart + 1, MyLength);
Window->ClustersFree += count;
//
// If this was not the last window this allocation spanned,
// advance to the next.
//
if (MyLength != count) {
Window++;
MyStart = Window->FirstCluster;
}
}
//
// Deallocation is now complete. Adjust the free cluster count.
//
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
}
#if DBG
if (Vcb->CurrentWindow->ClustersFree !=
RtlNumberOfClearBits(&Vcb->FreeClusterBitMap)) {
DbgPrint("%x vs %x\n", Vcb->CurrentWindow->ClustersFree,
RtlNumberOfClearBits(&Vcb->FreeClusterBitMap));
DbgPrint("%x for %x\n", ClusterIndex, ClusterCount);
}
#endif
FatUnlockFreeClusterBitMap( Vcb );
} _SEH2_FINALLY {
DebugUnwind( FatDeallocateDiskSpace );
//
// Is there any unwinding to do?
//
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
if ( _SEH2_AbnormalTermination() ) {
LBO LocalLbo;
VBO LocalVbo;
ULONG Index;
ULONG Clusters;
ULONG FatIndex;
ULONG PriorLastIndex;
//
// For each entry we already deallocated, reallocate it,
// chaining together as nessecary. Note that we continue
// up to and including the last "for" iteration even though
// the SetFatRun could not have been successful. This
// allows us a convienent way to re-link the final successful
// SetFatRun.
//
// It is possible that the reason we got here will prevent us
// from succeeding in this operation.
//
PriorLastIndex = 0;
for (Index = 0; Index <= McbIndex; Index++) {
FatGetNextMcbEntry(Vcb, Mcb, Index, &LocalVbo, &LocalLbo, &ByteCount);
if (ByteCount == 0xFFFFFFFF) {
//
// Special case the computation of ClusterCout
// when file is of max size (2^32 - 1).
//
Clusters = (1 << (32 - LogOfBytesPerCluster));
} else {
Clusters = ByteCount >> LogOfBytesPerCluster;
}
FatIndex = FatGetIndexFromLbo( Vcb, LocalLbo );
//
// We must always restore the prior iteration's last
// entry, pointing it to the first cluster of this run.
//
if (PriorLastIndex != 0) {
FatSetFatEntry( IrpContext,
Vcb,
PriorLastIndex,
(FAT_ENTRY)FatIndex );
}
//
// If this is not the last entry (the one that failed)
// then reallocate the disk space on the fat.
//
if ( Index < McbIndex ) {
FatAllocateClusters(IrpContext, Vcb, FatIndex, Clusters);
PriorLastIndex = FatIndex + Clusters - 1;
}
}
}
DebugTrace(-1, Dbg, "FatDeallocateDiskSpace -> (VOID)\n", 0);
} _SEH2_END;
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatSplitAllocation (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN OUT PLARGE_MCB Mcb,
IN VBO SplitAtVbo,
OUT PLARGE_MCB RemainingMcb
)
/*++
Routine Description:
This procedure takes a single mcb and splits its allocation into
two separate allocation units. The separation must only be done
on cluster boundaries, otherwise we bugcheck.
On the disk this actually works by inserting a FAT_CLUSTER_LAST into
the last index of the first part being split out.
Pictorially what is done is the following (where ! denotes the end of
the fat chain (i.e., FAT_CLUSTER_LAST)):
Mcb |--a--|--b--|--c--|--d--|--e--|--f--|
^
SplitAtVbo ---------------------+
RemainingMcb (empty)
becomes
Mcb |--a--|--b--|--c--!
RemainingMcb |--d--|--e--|--f--|
Arguments:
Vcb - Supplies the VCB being modified
Mcb - Supplies the MCB describing the allocation being split into
two parts. Upon return this Mcb now contains the first chain.
SplitAtVbo - Supplies the VBO of the first byte for the second chain
that we creating.
RemainingMcb - Receives the MCB describing the second chain of allocated
disk space. The caller passes in an initialized Mcb that
is filled in by this procedure STARTING AT VBO 0.
Return Value:
VOID - TRUE if the operation completed and FALSE if it had to
block but could not.
--*/
{
VBO SourceVbo;
VBO TargetVbo;
VBO DontCare;
LBO Lbo;
ULONG ByteCount;
#if DBG
ULONG BytesPerCluster;
#endif
PAGED_CODE();
DebugTrace(+1, Dbg, "FatSplitAllocation\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
DebugTrace( 0, Dbg, " SplitAtVbo = %8lx\n", SplitAtVbo);
DebugTrace( 0, Dbg, " RemainingMcb = %p\n", RemainingMcb);
#if DBG
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
#endif
//
// Assert that the split point is cluster alligned
//
NT_ASSERT( (SplitAtVbo & (BytesPerCluster - 1)) == 0 );
//
// We should never be handed an empty source MCB and asked to split
// at a non zero point.
//
NT_ASSERT( !((0 != SplitAtVbo) && (0 == FsRtlNumberOfRunsInLargeMcb( Mcb))));
//
// Assert we were given an empty target Mcb.
//
//
// This assert is commented out to avoid hitting in the Ea error
// path. In that case we will be using the same Mcb's to split the
// allocation that we used to merge them. The target Mcb will contain
// the runs that the split will attempt to insert.
//
//
// NT_ASSERT( FsRtlNumberOfRunsInMcb( RemainingMcb ) == 0 );
//
_SEH2_TRY {
//
// Move the runs after SplitAtVbo from the souce to the target
//
SourceVbo = SplitAtVbo;
TargetVbo = 0;
while (FatLookupMcbEntry(Vcb, Mcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
FatAddMcbEntry( Vcb, RemainingMcb, TargetVbo, Lbo, ByteCount );
FatRemoveMcbEntry( Vcb, Mcb, SourceVbo, ByteCount );
TargetVbo += ByteCount;
SourceVbo += ByteCount;
//
// If SourceVbo overflows, we were actually snipping off the end
// of the maximal file ... and are now done.
//
if (SourceVbo == 0) {
break;
}
}
//
// Mark the last pre-split cluster as a FAT_LAST_CLUSTER
//
if ( SplitAtVbo != 0 ) {
FatLookupLastMcbEntry( Vcb, Mcb, &DontCare, &Lbo, NULL );
FatSetFatEntry( IrpContext,
Vcb,
FatGetIndexFromLbo( Vcb, Lbo ),
FAT_CLUSTER_LAST );
}
} _SEH2_FINALLY {
DebugUnwind( FatSplitAllocation );
//
// If we got an exception, we must glue back together the Mcbs
//
if ( _SEH2_AbnormalTermination() ) {
TargetVbo = SplitAtVbo;
SourceVbo = 0;
while (FatLookupMcbEntry(Vcb, RemainingMcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
FatAddMcbEntry( Vcb, Mcb, TargetVbo, Lbo, ByteCount );
FatRemoveMcbEntry( Vcb, RemainingMcb, SourceVbo, ByteCount );
TargetVbo += ByteCount;
SourceVbo += ByteCount;
}
}
DebugTrace(-1, Dbg, "FatSplitAllocation -> (VOID)\n", 0);
} _SEH2_END;
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatMergeAllocation (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN OUT PLARGE_MCB Mcb,
IN PLARGE_MCB SecondMcb
)
/*++
Routine Description:
This routine takes two separate allocations described by two MCBs and
joins them together into one allocation.
Pictorially what is done is the following (where ! denotes the end of
the fat chain (i.e., FAT_CLUSTER_LAST)):
Mcb |--a--|--b--|--c--!
SecondMcb |--d--|--e--|--f--|
becomes
Mcb |--a--|--b--|--c--|--d--|--e--|--f--|
SecondMcb |--d--|--e--|--f--|
Arguments:
Vcb - Supplies the VCB being modified
Mcb - Supplies the MCB of the first allocation that is being modified.
Upon return this Mcb will also describe the newly enlarged
allocation
SecondMcb - Supplies the ZERO VBO BASED MCB of the second allocation
that is being appended to the first allocation. This
procedure leaves SecondMcb unchanged.
Return Value:
VOID - TRUE if the operation completed and FALSE if it had to
block but could not.
--*/
{
VBO SpliceVbo = 0;
LBO SpliceLbo;
VBO SourceVbo;
VBO TargetVbo = 0;
LBO Lbo;
ULONG ByteCount;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatMergeAllocation\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
DebugTrace( 0, Dbg, " SecondMcb = %p\n", SecondMcb);
_SEH2_TRY {
//
// Append the runs from SecondMcb to Mcb
//
(void)FatLookupLastMcbEntry( Vcb, Mcb, &SpliceVbo, &SpliceLbo, NULL );
SourceVbo = 0;
TargetVbo = SpliceVbo + 1;
while (FatLookupMcbEntry(Vcb, SecondMcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
FatAddMcbEntry( Vcb, Mcb, TargetVbo, Lbo, ByteCount );
SourceVbo += ByteCount;
TargetVbo += ByteCount;
}
//
// Link the last pre-merge cluster to the first cluster of SecondMcb
//
FatLookupMcbEntry( Vcb, SecondMcb, 0, &Lbo, (PULONG)NULL, NULL );
FatSetFatEntry( IrpContext,
Vcb,
FatGetIndexFromLbo( Vcb, SpliceLbo ),
(FAT_ENTRY)FatGetIndexFromLbo( Vcb, Lbo ) );
} _SEH2_FINALLY {
DebugUnwind( FatMergeAllocation );
//
// If we got an exception, we must remove the runs added to Mcb
//
if ( _SEH2_AbnormalTermination() ) {
ULONG CutLength;
if ((CutLength = TargetVbo - (SpliceVbo + 1)) != 0) {
FatRemoveMcbEntry( Vcb, Mcb, SpliceVbo + 1, CutLength);
}
}
DebugTrace(-1, Dbg, "FatMergeAllocation -> (VOID)\n", 0);
} _SEH2_END;
return;
}
//
// Internal support routine
//
CLUSTER_TYPE
FatInterpretClusterType (
IN PVCB Vcb,
IN FAT_ENTRY Entry
)
/*++
Routine Description:
This procedure tells the caller how to interpret the input fat table
entry. It will indicate if the fat cluster is available, resereved,
bad, the last one, or the another fat index. This procedure can deal
with both 12 and 16 bit fat.
Arguments:
Vcb - Supplies the Vcb to examine, yields 12/16 bit info
Entry - Supplies the fat entry to examine
Return Value:
CLUSTER_TYPE - Is the type of the input Fat entry
--*/
{
DebugTrace(+1, Dbg, "InterpretClusterType\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " Entry = %8lx\n", Entry);
PAGED_CODE();
switch(Vcb->AllocationSupport.FatIndexBitSize ) {
case 32:
Entry &= FAT32_ENTRY_MASK;
break;
case 12:
NT_ASSERT( Entry <= 0xfff );
if (Entry >= 0x0ff0) {
Entry |= 0x0FFFF000;
}
break;
default:
case 16:
NT_ASSERT( Entry <= 0xffff );
if (Entry >= 0x0fff0) {
Entry |= 0x0FFF0000;
}
break;
}
if (Entry == FAT_CLUSTER_AVAILABLE) {
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterAvailable\n", 0);
return FatClusterAvailable;
} else if (Entry < FAT_CLUSTER_RESERVED) {
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterNext\n", 0);
return FatClusterNext;
} else if (Entry < FAT_CLUSTER_BAD) {
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterReserved\n", 0);
return FatClusterReserved;
} else if (Entry == FAT_CLUSTER_BAD) {
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterBad\n", 0);
return FatClusterBad;
} else {
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterLast\n", 0);
return FatClusterLast;
}
}
//
// Internal support routine
//
VOID
FatLookupFatEntry (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG FatIndex,
IN OUT PULONG FatEntry,
IN OUT PFAT_ENUMERATION_CONTEXT Context
)
/*++
Routine Description:
This routine takes an index into the fat and gives back the value
in the Fat at this index. At any given time, for a 16 bit fat, this
routine allows only one page per volume of the fat to be pinned in
memory. For a 12 bit bit fat, the entire fat (max 6k) is pinned. This
extra layer of caching makes the vast majority of requests very
fast. The context for this caching stored in a structure in the Vcb.
Arguments:
Vcb - Supplies the Vcb to examine, yields 12/16 bit info,
fat access context, etc.
FatIndex - Supplies the fat index to examine.
FatEntry - Receives the fat entry pointed to by FatIndex. Note that
it must point to non-paged pool.
Context - This structure keeps track of a page of pinned fat between calls.
--*/
{
PAGED_CODE();
DebugTrace(+1, Dbg, "FatLookupFatEntry\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " FatIndex = %4x\n", FatIndex);
DebugTrace( 0, Dbg, " FatEntry = %8lx\n", FatEntry);
//
// Make sure they gave us a valid fat index.
//
FatVerifyIndexIsValid(IrpContext, Vcb, FatIndex);
//
// Case on 12 or 16 bit fats.
//
// In the 12 bit case (mostly floppies) we always have the whole fat
// (max 6k bytes) pinned during allocation operations. This is possibly
// a wee bit slower, but saves headaches over fat entries with 8 bits
// on one page, and 4 bits on the next.
//
// The 16 bit case always keeps the last used page pinned until all
// operations are done and it is unpinned.
//
//
// DEAL WITH 12 BIT CASE
//
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
//
// Check to see if the fat is already pinned, otherwise pin it.
//
if (Context->Bcb == NULL) {
FatReadVolumeFile( IrpContext,
Vcb,
FatReservedBytes( &Vcb->Bpb ),
FatBytesPerFat( &Vcb->Bpb ),
&Context->Bcb,
&Context->PinnedPage );
}
//
// Load the return value.
//
FatLookup12BitEntry( Context->PinnedPage, FatIndex, FatEntry );
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
//
// DEAL WITH 32 BIT CASE
//
ULONG PageEntryOffset;
ULONG OffsetIntoVolumeFile;
//
// Initialize two local variables that help us.
//
OffsetIntoVolumeFile = FatReservedBytes(&Vcb->Bpb) + FatIndex * sizeof(FAT_ENTRY);
PageEntryOffset = (OffsetIntoVolumeFile % PAGE_SIZE) / sizeof(FAT_ENTRY);
//
// Check to see if we need to read in a new page of fat
//
if ((Context->Bcb == NULL) ||
(OffsetIntoVolumeFile / PAGE_SIZE != Context->VboOfPinnedPage / PAGE_SIZE)) {
//
// The entry wasn't in the pinned page, so must we unpin the current
// page (if any) and read in a new page.
//
FatUnpinBcb( IrpContext, Context->Bcb );
FatReadVolumeFile( IrpContext,
Vcb,
OffsetIntoVolumeFile & ~(PAGE_SIZE - 1),
PAGE_SIZE,
&Context->Bcb,
&Context->PinnedPage );
Context->VboOfPinnedPage = OffsetIntoVolumeFile & ~(PAGE_SIZE - 1);
}
//
// Grab the fat entry from the pinned page, and return
//
*FatEntry = ((PULONG)(Context->PinnedPage))[PageEntryOffset] & FAT32_ENTRY_MASK;
} else {
//
// DEAL WITH 16 BIT CASE
//
ULONG PageEntryOffset;
ULONG OffsetIntoVolumeFile;
//
// Initialize two local variables that help us.
//
OffsetIntoVolumeFile = FatReservedBytes(&Vcb->Bpb) + FatIndex * sizeof(USHORT);
PageEntryOffset = (OffsetIntoVolumeFile % PAGE_SIZE) / sizeof(USHORT);
//
// Check to see if we need to read in a new page of fat
//
if ((Context->Bcb == NULL) ||
(OffsetIntoVolumeFile / PAGE_SIZE != Context->VboOfPinnedPage / PAGE_SIZE)) {
//
// The entry wasn't in the pinned page, so must we unpin the current
// page (if any) and read in a new page.
//
FatUnpinBcb( IrpContext, Context->Bcb );
FatReadVolumeFile( IrpContext,
Vcb,
OffsetIntoVolumeFile & ~(PAGE_SIZE - 1),
PAGE_SIZE,
&Context->Bcb,
&Context->PinnedPage );
Context->VboOfPinnedPage = OffsetIntoVolumeFile & ~(PAGE_SIZE - 1);
}
//
// Grab the fat entry from the pinned page, and return
//
*FatEntry = ((PUSHORT)(Context->PinnedPage))[PageEntryOffset];
}
DebugTrace(-1, Dbg, "FatLookupFatEntry -> (VOID)\n", 0);
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatSetFatEntry (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG FatIndex,
IN FAT_ENTRY FatEntry
)
/*++
Routine Description:
This routine takes an index into the fat and puts a value in the Fat
at this index. The routine special cases 12, 16 and 32 bit fats. In
all cases we go to the cache manager for a piece of the fat.
We have a special form of this call for setting the DOS-style dirty bit.
Unlike the dirty bit in the boot sector, we do not go to special effort
to make sure that this hits the disk synchronously - if the system goes
down in the window between the dirty bit being set in the boot sector
and the FAT index zero dirty bit being lazy written, then life is tough.
The only possible scenario is that Win9x may see what it thinks is a clean
volume that really isn't (hopefully Memphis will pay attention to our dirty
bit as well). The dirty bit will get out quickly, and if heavy activity is
occurring, then the dirty bit should actually be there virtually all of the
time since the act of cleaning the volume is the "rare" occurance.
There are synchronization concerns that would crop up if we tried to make
this synchronous. This thread may already own the Bcb shared for the first
sector of the FAT (so we can't get it exclusive for a writethrough). This
would require some more serious replumbing to work around than I want to
consider at this time.
We can and do, however, synchronously set the bit clean.
At this point the reader should understand why the NT dirty bit is where it is.
Arguments:
Vcb - Supplies the Vcb to examine, yields 12/16/32 bit info, etc.
FatIndex - Supplies the destination fat index.
FatEntry - Supplies the source fat entry.
--*/
{
LBO Lbo;
PBCB Bcb = NULL;
ULONG SectorSize;
ULONG OffsetIntoVolumeFile;
ULONG WasWait = TRUE;
BOOLEAN RegularOperation = TRUE;
BOOLEAN CleaningOperation = FALSE;
BOOLEAN ReleaseMutex = FALSE;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatSetFatEntry\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " FatIndex = %4x\n", FatIndex);
DebugTrace( 0, Dbg, " FatEntry = %4x\n", FatEntry);
//
// Make sure they gave us a valid fat index if this isn't the special
// clean-bit modifying call.
//
if (FatIndex == FAT_DIRTY_BIT_INDEX) {
//
// We are setting the clean bit state. Of course, we could
// have corruption that would cause us to try to fiddle the
// reserved index - we guard against this by having the
// special entry values use the reserved high 4 bits that
// we know that we'll never try to set.
//
//
// We don't want to repin the FAT pages involved here. Just
// let the lazy writer hit them when it can.
//
RegularOperation = FALSE;
switch (FatEntry) {
case FAT_CLEAN_VOLUME:
FatEntry = (FAT_ENTRY)FAT_CLEAN_ENTRY;
CleaningOperation = TRUE;
break;
case FAT_DIRTY_VOLUME:
switch (Vcb->AllocationSupport.FatIndexBitSize) {
case 12:
FatEntry = FAT12_DIRTY_ENTRY;
break;
case 32:
FatEntry = FAT32_DIRTY_ENTRY;
break;
default:
FatEntry = FAT16_DIRTY_ENTRY;
break;
}
break;
default:
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
break;
}
//
// Disable dirtying semantics for the duration of this operation. Force this
// operation to wait for the duration.
//
WasWait = FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT );
SetFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT | IRP_CONTEXT_FLAG_DISABLE_DIRTY );
} else {
NT_ASSERT( !(FatEntry & ~FAT32_ENTRY_MASK) );
FatVerifyIndexIsValid(IrpContext, Vcb, FatIndex);
}
//
// Set Sector Size
//
SectorSize = 1 << Vcb->AllocationSupport.LogOfBytesPerSector;
//
// Case on 12 or 16 bit fats.
//
// In the 12 bit case (mostly floppies) we always have the whole fat
// (max 6k bytes) pinned during allocation operations. This is possibly
// a wee bit slower, but saves headaches over fat entries with 8 bits
// on one page, and 4 bits on the next.
//
// In the 16 bit case we only read the page that we need to set the fat
// entry.
//
//
// DEAL WITH 12 BIT CASE
//
_SEH2_TRY {
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
PVOID PinnedFat;
//
// Make sure we have a valid entry
//
FatEntry &= 0xfff;
//
// We read in the entire fat. Note that using prepare write marks
// the bcb pre-dirty, so we don't have to do it explicitly.
//
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) + FatIndex * 3 / 2;
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
FatReservedBytes( &Vcb->Bpb ),
FatBytesPerFat( &Vcb->Bpb ),
&Bcb,
&PinnedFat,
RegularOperation,
FALSE );
//
// Mark the sector(s) dirty in the DirtyFatMcb. This call is
// complicated somewhat for the 12 bit case since a single
// entry write can span two sectors (and pages).
//
// Get the Lbo for the sector where the entry starts, and add it to
// the dirty fat Mcb.
//
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
//
// If the entry started on the last byte of the sector, it continues
// to the next sector, so mark the next sector dirty as well.
//
// Note that this entry will simply coalese with the last entry,
// so this operation cannot fail. Also if we get this far, we have
// made it, so no unwinding will be needed.
//
if ( (OffsetIntoVolumeFile & (SectorSize - 1)) == (SectorSize - 1) ) {
Lbo += SectorSize;
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
}
//
// Store the entry into the fat; we need a little synchonization
// here and can't use a spinlock since the bytes might not be
// resident.
//
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
FatSet12BitEntry( PinnedFat, FatIndex, FatEntry );
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
//
// DEAL WITH 32 BIT CASE
//
PULONG PinnedFatEntry32;
//
// Read in a new page of fat
//
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) +
FatIndex * sizeof( FAT_ENTRY );
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
OffsetIntoVolumeFile,
sizeof(FAT_ENTRY),
&Bcb,
(PVOID *)&PinnedFatEntry32,
RegularOperation,
FALSE );
//
// Mark the sector dirty in the DirtyFatMcb
//
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
//
// Store the FatEntry to the pinned page.
//
// Preserve the reserved bits in FAT32 entries in the file heap.
//
#ifdef ALPHA
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
#endif // ALPHA
if (FatIndex != FAT_DIRTY_BIT_INDEX) {
*PinnedFatEntry32 = ((*PinnedFatEntry32 & ~FAT32_ENTRY_MASK) | FatEntry);
} else {
*PinnedFatEntry32 = FatEntry;
}
#ifdef ALPHA
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
#endif // ALPHA
} else {
//
// DEAL WITH 16 BIT CASE
//
PUSHORT PinnedFatEntry;
//
// Read in a new page of fat
//
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) +
FatIndex * sizeof(USHORT);
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
OffsetIntoVolumeFile,
sizeof(USHORT),
&Bcb,
(PVOID *)&PinnedFatEntry,
RegularOperation,
FALSE );
//
// Mark the sector dirty in the DirtyFatMcb
//
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
//
// Store the FatEntry to the pinned page.
//
// We need extra synchronization here for broken architectures
// like the ALPHA that don't support atomic 16 bit writes.
//
#ifdef ALPHA
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
#endif // ALPHA
*PinnedFatEntry = (USHORT)FatEntry;
#ifdef ALPHA
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
#endif // ALPHA
}
} _SEH2_FINALLY {
DebugUnwind( FatSetFatEntry );
//
// Re-enable volume dirtying in case this was a dirty bit operation.
//
ClearFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_DIRTY );
//
// Make this operation asynchronous again if needed.
//
if (!WasWait) {
ClearFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT );
}
//
// If we still somehow have the Mutex, release it.
//
if (ReleaseMutex) {
NT_ASSERT( _SEH2_AbnormalTermination() );
FatUnlockFreeClusterBitMap( Vcb );
}
//
// Unpin the Bcb. For cleaning operations or if the corruption was detected while mounting we make this write-through.
//
if ((CleaningOperation ||
FlagOn(Vcb->VcbState, VCB_STATE_FLAG_MOUNT_IN_PROGRESS)) &&
Bcb) {
IO_STATUS_BLOCK IgnoreStatus;
CcRepinBcb( Bcb );
CcUnpinData( Bcb );
DbgDoit( IrpContext->PinCount -= 1 );
CcUnpinRepinnedBcb( Bcb, TRUE, &IgnoreStatus );
} else {
FatUnpinBcb(IrpContext, Bcb);
}
DebugTrace(-1, Dbg, "FatSetFatEntry -> (VOID)\n", 0);
} _SEH2_END;
return;
}
//
// Internal support routine
//
VOID
FatSetFatRun (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG StartingFatIndex,
IN ULONG ClusterCount,
IN BOOLEAN ChainTogether
)
/*++
Routine Description:
This routine sets a continuous run of clusters in the fat. If ChainTogether
is TRUE, then the clusters are linked together as in normal Fat fasion,
with the last cluster receiving FAT_CLUSTER_LAST. If ChainTogether is
FALSE, all the entries are set to FAT_CLUSTER_AVAILABLE, effectively
freeing all the clusters in the run.
Arguments:
Vcb - Supplies the Vcb to examine, yields 12/16 bit info, etc.
StartingFatIndex - Supplies the destination fat index.
ClusterCount - Supplies the number of contiguous clusters to work on.
ChainTogether - Tells us whether to fill the entries with links, or
FAT_CLUSTER_AVAILABLE
Return Value:
VOID
--*/
{
#define MAXCOUNTCLUS 0x10000
#define COUNTSAVEDBCBS ((MAXCOUNTCLUS * sizeof(FAT_ENTRY) / PAGE_SIZE) + 2)
PBCB SavedBcbs[COUNTSAVEDBCBS][2];
ULONG SectorSize;
ULONG Cluster;
LBO StartSectorLbo;
LBO FinalSectorLbo;
LBO Lbo;
PVOID PinnedFat;
BOOLEAN ReleaseMutex = FALSE;
ULONG SavedStartingFatIndex = StartingFatIndex;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatSetFatRun\n", 0);
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, " StartingFatIndex = %8x\n", StartingFatIndex);
DebugTrace( 0, Dbg, " ClusterCount = %8lx\n", ClusterCount);
DebugTrace( 0, Dbg, " ChainTogether = %s\n", ChainTogether ? "TRUE":"FALSE");
//
// Make sure they gave us a valid fat run.
//
FatVerifyIndexIsValid(IrpContext, Vcb, StartingFatIndex);
FatVerifyIndexIsValid(IrpContext, Vcb, StartingFatIndex + ClusterCount - 1);
//
// Check special case
//
if (ClusterCount == 0) {
DebugTrace(-1, Dbg, "FatSetFatRun -> (VOID)\n", 0);
return;
}
//
// Set Sector Size
//
SectorSize = 1 << Vcb->AllocationSupport.LogOfBytesPerSector;
//
// Case on 12 or 16 bit fats.
//
// In the 12 bit case (mostly floppies) we always have the whole fat
// (max 6k bytes) pinned during allocation operations. This is possibly
// a wee bit slower, but saves headaches over fat entries with 8 bits
// on one page, and 4 bits on the next.
//
// In the 16 bit case we only read one page at a time, as needed.
//
//
// DEAL WITH 12 BIT CASE
//
_SEH2_TRY {
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
//
// We read in the entire fat. Note that using prepare write marks
// the bcb pre-dirty, so we don't have to do it explicitly.
//
RtlZeroMemory( &SavedBcbs[0][0], 2 * sizeof(PBCB) * 2);
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
FatReservedBytes( &Vcb->Bpb ),
FatBytesPerFat( &Vcb->Bpb ),
&SavedBcbs[0][0],
&PinnedFat,
TRUE,
FALSE );
//
// Mark the affected sectors dirty. Note that FinalSectorLbo is
// the Lbo of the END of the entry (Thus * 3 + 2). This makes sure
// we catch the case of a dirty fat entry straddling a sector boundry.
//
// Note that if the first AddMcbEntry succeeds, all following ones
// will simply coalese, and thus also succeed.
//
StartSectorLbo = (FatReservedBytes( &Vcb->Bpb ) + StartingFatIndex * 3 / 2)
& ~(SectorSize - 1);
FinalSectorLbo = (FatReservedBytes( &Vcb->Bpb ) + ((StartingFatIndex +
ClusterCount) * 3 + 2) / 2) & ~(SectorSize - 1);
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
}
//
// Store the entries into the fat; we need a little
// synchonization here and can't use a spinlock since the bytes
// might not be resident.
//
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
for (Cluster = StartingFatIndex;
Cluster < StartingFatIndex + ClusterCount - 1;
Cluster++) {
FatSet12BitEntry( PinnedFat,
Cluster,
ChainTogether ? Cluster + 1 : FAT_CLUSTER_AVAILABLE );
}
//
// Save the last entry
//
FatSet12BitEntry( PinnedFat,
Cluster,
ChainTogether ?
FAT_CLUSTER_LAST & 0xfff : FAT_CLUSTER_AVAILABLE );
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
//
// DEAL WITH 32 BIT CASE
//
for (;;) {
VBO StartOffsetInVolume;
VBO FinalOffsetInVolume;
ULONG Page;
ULONG FinalCluster;
PULONG FatEntry = NULL;
ULONG ClusterCountThisRun;
StartOffsetInVolume = FatReservedBytes(&Vcb->Bpb) +
StartingFatIndex * sizeof(FAT_ENTRY);
if (ClusterCount > MAXCOUNTCLUS) {
ClusterCountThisRun = MAXCOUNTCLUS;
} else {
ClusterCountThisRun = ClusterCount;
}
FinalOffsetInVolume = StartOffsetInVolume +
(ClusterCountThisRun - 1) * sizeof(FAT_ENTRY);
{
ULONG NumberOfPages;
ULONG Offset;
NumberOfPages = (FinalOffsetInVolume / PAGE_SIZE) -
(StartOffsetInVolume / PAGE_SIZE) + 1;
RtlZeroMemory( &SavedBcbs[0][0], (NumberOfPages + 1) * sizeof(PBCB) * 2 );
for ( Page = 0, Offset = StartOffsetInVolume & ~(PAGE_SIZE - 1);
Page < NumberOfPages;
Page++, Offset += PAGE_SIZE ) {
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
Offset,
PAGE_SIZE,
&SavedBcbs[Page][0],
(PVOID *)&SavedBcbs[Page][1],
TRUE,
FALSE );
if (Page == 0) {
FatEntry = (PULONG)((PUCHAR)SavedBcbs[0][1] +
(StartOffsetInVolume % PAGE_SIZE));
}
}
}
//
// Mark the run dirty
//
StartSectorLbo = StartOffsetInVolume & ~(SectorSize - 1);
FinalSectorLbo = FinalOffsetInVolume & ~(SectorSize - 1);
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO)Lbo, Lbo, SectorSize );
}
//
// Store the entries
//
// We need extra synchronization here for broken architectures
// like the ALPHA that don't support atomic 16 bit writes.
//
#ifdef ALPHA
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
#endif // ALPHA
FinalCluster = StartingFatIndex + ClusterCountThisRun - 1;
Page = 0;
for (Cluster = StartingFatIndex;
Cluster <= FinalCluster;
Cluster++, FatEntry++) {
//
// If we just crossed a page boundry (as opposed to starting
// on one), update our idea of FatEntry.
if ( (((ULONG_PTR)FatEntry & (PAGE_SIZE-1)) == 0) &&
(Cluster != StartingFatIndex) ) {
Page += 1;
FatEntry = (PULONG)SavedBcbs[Page][1];
}
*FatEntry = ChainTogether ? (FAT_ENTRY)(Cluster + 1) :
FAT_CLUSTER_AVAILABLE;
}
//
// Fix up the last entry if we were chaining together
//
if ((ClusterCount <= MAXCOUNTCLUS) &&
ChainTogether ) {
*(FatEntry-1) = FAT_CLUSTER_LAST;
}
#ifdef ALPHA
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
#endif // ALPHA
{
ULONG i;
//
// Unpin the Bcbs
//
for (i = 0; (i < COUNTSAVEDBCBS) && (SavedBcbs[i][0] != NULL); i++) {
FatUnpinBcb( IrpContext, SavedBcbs[i][0] );
SavedBcbs[i][0] = NULL;
}
}
if (ClusterCount <= MAXCOUNTCLUS) {
break;
} else {
StartingFatIndex += MAXCOUNTCLUS;
ClusterCount -= MAXCOUNTCLUS;
}
}
} else {
//
// DEAL WITH 16 BIT CASE
//
VBO StartOffsetInVolume;
VBO FinalOffsetInVolume;
ULONG Page;
ULONG FinalCluster;
PUSHORT FatEntry = NULL;
StartOffsetInVolume = FatReservedBytes(&Vcb->Bpb) +
StartingFatIndex * sizeof(USHORT);
FinalOffsetInVolume = StartOffsetInVolume +
(ClusterCount - 1) * sizeof(USHORT);
//
// Read in one page of fat at a time. We cannot read in the
// all of the fat we need because of cache manager limitations.
//
// SavedBcb was initialized to be able to hold the largest
// possible number of pages in a fat plus and extra one to
// accomadate the boot sector, plus one more to make sure there
// is enough room for the RtlZeroMemory below that needs the mark
// the first Bcb after all the ones we will use as an end marker.
//
{
ULONG NumberOfPages;
ULONG Offset;
NumberOfPages = (FinalOffsetInVolume / PAGE_SIZE) -
(StartOffsetInVolume / PAGE_SIZE) + 1;
RtlZeroMemory( &SavedBcbs[0][0], (NumberOfPages + 1) * sizeof(PBCB) * 2 );
for ( Page = 0, Offset = StartOffsetInVolume & ~(PAGE_SIZE - 1);
Page < NumberOfPages;
Page++, Offset += PAGE_SIZE ) {
FatPrepareWriteVolumeFile( IrpContext,
Vcb,
Offset,
PAGE_SIZE,
&SavedBcbs[Page][0],
(PVOID *)&SavedBcbs[Page][1],
TRUE,
FALSE );
if (Page == 0) {
FatEntry = (PUSHORT)((PUCHAR)SavedBcbs[0][1] +
(StartOffsetInVolume % PAGE_SIZE));
}
}
}
//
// Mark the run dirty
//
StartSectorLbo = StartOffsetInVolume & ~(SectorSize - 1);
FinalSectorLbo = FinalOffsetInVolume & ~(SectorSize - 1);
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
}
//
// Store the entries
//
// We need extra synchronization here for broken architectures
// like the ALPHA that don't support atomic 16 bit writes.
//
#ifdef ALPHA
FatLockFreeClusterBitMap( Vcb );
ReleaseMutex = TRUE;
#endif // ALPHA
FinalCluster = StartingFatIndex + ClusterCount - 1;
Page = 0;
for (Cluster = StartingFatIndex;
Cluster <= FinalCluster;
Cluster++, FatEntry++) {
//
// If we just crossed a page boundry (as opposed to starting
// on one), update our idea of FatEntry.
if ( (((ULONG_PTR)FatEntry & (PAGE_SIZE-1)) == 0) &&
(Cluster != StartingFatIndex) ) {
Page += 1;
FatEntry = (PUSHORT)SavedBcbs[Page][1];
}
*FatEntry = (USHORT) (ChainTogether ? (FAT_ENTRY)(Cluster + 1) :
FAT_CLUSTER_AVAILABLE);
}
//
// Fix up the last entry if we were chaining together
//
if ( ChainTogether ) {
#ifdef _MSC_VER
#pragma warning( suppress: 4310 )
#endif
*(FatEntry-1) = (USHORT)FAT_CLUSTER_LAST;
}
#ifdef ALPHA
FatUnlockFreeClusterBitMap( Vcb );
ReleaseMutex = FALSE;
#endif // ALPHA
}
} _SEH2_FINALLY {
ULONG i;
DebugUnwind( FatSetFatRun );
//
// If we still somehow have the Mutex, release it.
//
if (ReleaseMutex) {
NT_ASSERT( _SEH2_AbnormalTermination() );
FatUnlockFreeClusterBitMap( Vcb );
}
//
// Unpin the Bcbs
//
for (i = 0; (i < COUNTSAVEDBCBS) && (SavedBcbs[i][0] != NULL); i++) {
FatUnpinBcb( IrpContext, SavedBcbs[i][0] );
}
//
// At this point nothing in this finally clause should have raised.
// So, now comes the unsafe (sigh) stuff.
//
if ( _SEH2_AbnormalTermination() &&
(Vcb->AllocationSupport.FatIndexBitSize == 32) ) {
//
// Fat32 unwind
//
// This case is more complex because the FAT12 and FAT16 cases
// pin all the needed FAT pages (128K max), after which it
// can't fail, before changing any FAT entries. In the Fat32
// case, it may not be practical to pin all the needed FAT
// pages, because that could span many megabytes. So Fat32
// attacks in chunks, and if a failure occurs once the first
// chunk has been updated, we have to back out the updates.
//
// The unwind consists of walking back over each FAT entry we
// have changed, setting it back to the previous value. Note
// that the previous value with either be FAT_CLUSTER_AVAILABLE
// (if ChainTogether==TRUE) or a simple link to the successor
// (if ChainTogether==FALSE).
//
// We concede that any one of these calls could fail too; our
// objective is to make this case no more likely than the case
// for a file consisting of multiple disjoint runs.
//
while ( StartingFatIndex > SavedStartingFatIndex ) {
StartingFatIndex--;
FatSetFatEntry( IrpContext, Vcb, StartingFatIndex,
ChainTogether ?
StartingFatIndex + 1 : FAT_CLUSTER_AVAILABLE );
}
}
DebugTrace(-1, Dbg, "FatSetFatRun -> (VOID)\n", 0);
} _SEH2_END;
return;
}
//
// Internal support routine
//
UCHAR
FatLogOf (
IN ULONG Value
)
/*++
Routine Description:
This routine just computes the base 2 log of an integer. It is only used
on objects that are know to be powers of two.
Arguments:
Value - The value to take the base 2 log of.
Return Value:
UCHAR - The base 2 log of Value.
--*/
{
UCHAR Log = 0;
#if FASTFATDBG
ULONG OrigValue = Value;
#endif
PAGED_CODE();
//
// Knock bits off until we we get a one at position 0
//
while ( (Value & 0xfffffffe) != 0 ) {
Log++;
Value >>= 1;
}
//
// If there was more than one bit set, the file system messed up,
// Bug Check.
//
if (Value != 0x1) {
DebugTrace(+1, Dbg, "LogOf\n", 0);
DebugTrace( 0, Dbg, " Value = %8lx\n", OrigValue);
DebugTrace( 0, Dbg, "Received non power of 2.\n", 0);
DebugTrace(-1, Dbg, "LogOf -> %8lx\n", Log);
#ifdef _MSC_VER
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
#endif
FatBugCheck( Value, Log, 0 );
}
return Log;
}
VOID
FatExamineFatEntries(
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN ULONG StartIndex OPTIONAL,
IN ULONG EndIndex OPTIONAL,
IN BOOLEAN SetupWindows,
IN PFAT_WINDOW SwitchToWindow OPTIONAL,
IN PULONG BitMapBuffer OPTIONAL
)
/*++
Routine Description:
This routine handles scanning a segment of the FAT into in-memory structures.
There are three fundamental cases, with variations depending on the FAT type:
1) During volume setup, FatSetupAllocations
1a) for FAT12/16, read the FAT into our free clusterbitmap
1b) for FAT32, perform the initial scan for window free cluster counts
2) Switching FAT32 windows on the fly during system operation
3) Reading arbitrary segments of the FAT for the purposes of the GetVolumeBitmap
call (only for FAT32)
There really is too much going on in here. At some point this should be
substantially rewritten.
Arguments:
Vcb - Supplies the volume involved
StartIndex - Supplies the starting cluster, ignored if SwitchToWindow supplied
EndIndex - Supplies the ending cluster, ignored if SwitchToWindow supplied
SetupWindows - Indicates if we are doing the initial FAT32 scan
SwitchToWindow - Supplies the FAT window we are examining and will switch to
BitMapBuffer - Supplies a specific bitmap to fill in, if not supplied we fill
in the volume free cluster bitmap if !SetupWindows
Return Value:
None. Lots of side effects.
--*/
{
ULONG FatIndexBitSize;
ULONG Page = 0;
ULONG Offset = 0;
ULONG FatIndex;
FAT_ENTRY FatEntry = FAT_CLUSTER_AVAILABLE;
FAT_ENTRY FirstFatEntry = FAT_CLUSTER_AVAILABLE;
PUSHORT FatBuffer;
PVOID pv;
PBCB Bcb = NULL;
ULONG EntriesPerWindow;
ULONG ClustersThisRun;
ULONG StartIndexOfThisRun;
PULONG FreeClusterCount = NULL;
PFAT_WINDOW CurrentWindow = NULL;
PVOID NewBitMapBuffer = NULL;
PRTL_BITMAP BitMap = NULL;
RTL_BITMAP PrivateBitMap;
ULONG ClusterSize = 0;
ULONG PrefetchPages = 0;
ULONG FatPages = 0;
VBO BadClusterVbo = 0;
LBO Lbo = 0;
enum RunType {
FreeClusters,
AllocatedClusters,
UnknownClusters
} CurrentRun;
PAGED_CODE();
//
// Now assert correct usage.
//
FatIndexBitSize = Vcb->AllocationSupport.FatIndexBitSize;
NT_ASSERT( !(SetupWindows && (SwitchToWindow || BitMapBuffer)));
NT_ASSERT( !(SetupWindows && FatIndexBitSize != 32));
if (Vcb->NumberOfWindows > 1) {
//
// FAT32: Calculate the number of FAT entries covered by a window. This is
// equal to the number of bits in the freespace bitmap, the size of which
// is hardcoded.
//
EntriesPerWindow = MAX_CLUSTER_BITMAP_SIZE;
} else {
EntriesPerWindow = Vcb->AllocationSupport.NumberOfClusters;
}
//
// We will also fill in the cumulative count of free clusters for
// the entire volume. If this is not appropriate, NULL it out
// shortly.
//
FreeClusterCount = &Vcb->AllocationSupport.NumberOfFreeClusters;
if (SetupWindows) {
NT_ASSERT(BitMapBuffer == NULL);
//
// In this case we're just supposed to scan the fat and set up
// the information regarding where the buckets fall and how many
// free clusters are in each.
//
// It is fine to monkey with the real windows, we must be able
// to do this to activate the volume.
//
BitMap = NULL;
CurrentWindow = &Vcb->Windows[0];
CurrentWindow->FirstCluster = StartIndex;
CurrentWindow->ClustersFree = 0;
//
// We always wish to calculate total free clusters when
// setting up the FAT windows.
//
} else if (BitMapBuffer == NULL) {
//
// We will be filling in the free cluster bitmap for the volume.
// Careful, we can raise out of here and be hopelessly hosed if
// we built this up in the main bitmap/window itself.
//
// For simplicity's sake, we'll do the swap for everyone. FAT32
// provokes the need since we can't tolerate partial results
// when switching windows.
//
NT_ASSERT( SwitchToWindow );
CurrentWindow = SwitchToWindow;
StartIndex = CurrentWindow->FirstCluster;
EndIndex = CurrentWindow->LastCluster;
BitMap = &PrivateBitMap;
NewBitMapBuffer = FsRtlAllocatePoolWithTag( PagedPool,
(EntriesPerWindow + 7) / 8,
TAG_FAT_BITMAP );
RtlInitializeBitMap( &PrivateBitMap,
NewBitMapBuffer,
EndIndex - StartIndex + 1);
if ((FatIndexBitSize == 32) &&
(Vcb->NumberOfWindows > 1)) {
//
// We do not wish count total clusters here.
//
FreeClusterCount = NULL;
}
} else {
BitMap = &PrivateBitMap;
RtlInitializeBitMap(&PrivateBitMap,
BitMapBuffer,
EndIndex - StartIndex + 1);
//
// We do not count total clusters here.
//
FreeClusterCount = NULL;
}
//
// Now, our start index better be in the file heap.
//
NT_ASSERT( StartIndex >= 2 );
_SEH2_TRY {
//
// Pick up the initial chunk of the FAT and first entry.
//
if (FatIndexBitSize == 12) {
//
// We read in the entire fat in the 12 bit case.
//
FatReadVolumeFile( IrpContext,
Vcb,
FatReservedBytes( &Vcb->Bpb ),
FatBytesPerFat( &Vcb->Bpb ),
&Bcb,
(PVOID *)&FatBuffer );
FatLookup12BitEntry(FatBuffer, 0, &FirstFatEntry);
} else {
//
// Read in one page of fat at a time. We cannot read in the
// all of the fat we need because of cache manager limitations.
//
ULONG BytesPerEntry = FatIndexBitSize >> 3;
FatPages = (FatReservedBytes(&Vcb->Bpb) + FatBytesPerFat(&Vcb->Bpb) + (PAGE_SIZE - 1)) / PAGE_SIZE;
Page = (FatReservedBytes(&Vcb->Bpb) + StartIndex * BytesPerEntry) / PAGE_SIZE;
Offset = Page * PAGE_SIZE;
//
// Prefetch the FAT entries in memory for optimal performance.
//
PrefetchPages = FatPages - Page;
if (PrefetchPages > FAT_PREFETCH_PAGE_COUNT) {
PrefetchPages = ALIGN_UP_BY(Page, FAT_PREFETCH_PAGE_COUNT) - Page;
}
#if (NTDDI_VERSION >= NTDDI_WIN8)
FatPrefetchPages( IrpContext,
Vcb->VirtualVolumeFile,
Page,
PrefetchPages );
#endif
FatReadVolumeFile( IrpContext,
Vcb,
Offset,
PAGE_SIZE,
&Bcb,
&pv);
if (FatIndexBitSize == 32) {
FatBuffer = (PUSHORT)((PUCHAR)pv +
(FatReservedBytes(&Vcb->Bpb) + StartIndex * BytesPerEntry) %
PAGE_SIZE);
FirstFatEntry = *((PULONG)FatBuffer);
FirstFatEntry = FirstFatEntry & FAT32_ENTRY_MASK;
} else {
FatBuffer = (PUSHORT)((PUCHAR)pv +
FatReservedBytes(&Vcb->Bpb) % PAGE_SIZE) + 2;
FirstFatEntry = *FatBuffer;
}
}
ClusterSize = 1 << (Vcb->AllocationSupport.LogOfBytesPerCluster);
CurrentRun = (FirstFatEntry == FAT_CLUSTER_AVAILABLE) ?
FreeClusters : AllocatedClusters;
StartIndexOfThisRun = StartIndex;
for (FatIndex = StartIndex; FatIndex <= EndIndex; FatIndex++) {
if (FatIndexBitSize == 12) {
FatLookup12BitEntry(FatBuffer, FatIndex, &FatEntry);
} else {
//
// If we are setting up the FAT32 windows and have stepped into a new
// bucket, finalize this one and move forward.
//
if (SetupWindows &&
FatIndex > StartIndex &&
(FatIndex - 2) % EntriesPerWindow == 0) {
CurrentWindow->LastCluster = FatIndex - 1;
if (CurrentRun == FreeClusters) {
//
// We must be counting clusters in order to modify the
// contents of the window.
//
NT_ASSERT( FreeClusterCount );
ClustersThisRun = FatIndex - StartIndexOfThisRun;
CurrentWindow->ClustersFree += ClustersThisRun;
if (FreeClusterCount) {
*FreeClusterCount += ClustersThisRun;
}
} else {
NT_ASSERT(CurrentRun == AllocatedClusters);
}
StartIndexOfThisRun = FatIndex;
CurrentRun = UnknownClusters;
CurrentWindow++;
CurrentWindow->ClustersFree = 0;
CurrentWindow->FirstCluster = FatIndex;
}
//
// If we just stepped onto a new page, grab a new pointer.
//
if (((ULONG_PTR)FatBuffer & (PAGE_SIZE - 1)) == 0) {
FatUnpinBcb( IrpContext, Bcb );
Page++;
Offset += PAGE_SIZE;
#if (NTDDI_VERSION >= NTDDI_WIN8)
//
// If we have exhausted all the prefetch pages, prefetch the next chunk.
//
if (--PrefetchPages == 0) {
PrefetchPages = FatPages - Page;
if (PrefetchPages > FAT_PREFETCH_PAGE_COUNT) {
PrefetchPages = FAT_PREFETCH_PAGE_COUNT;
}
FatPrefetchPages( IrpContext,
Vcb->VirtualVolumeFile,
Page,
PrefetchPages );
}
#endif
FatReadVolumeFile( IrpContext,
Vcb,
Offset,
PAGE_SIZE,
&Bcb,
&pv );
FatBuffer = (PUSHORT)pv;
}
if (FatIndexBitSize == 32) {
#ifndef __REACTOS__
#ifdef _MSC_VER
#pragma warning( suppress: 4213 )
#endif
FatEntry = *((PULONG)FatBuffer)++;
FatEntry = FatEntry & FAT32_ENTRY_MASK;
#else
FatEntry = *((PULONG)FatBuffer);
FatBuffer += 2; /* PUSHORT FatBuffer */
FatEntry = FatEntry & FAT32_ENTRY_MASK;
#endif
} else {
FatEntry = *FatBuffer;
FatBuffer += 1;
}
}
if (CurrentRun == UnknownClusters) {
CurrentRun = (FatEntry == FAT_CLUSTER_AVAILABLE) ?
FreeClusters : AllocatedClusters;
}
//
// Are we switching from a free run to an allocated run?
//
if (CurrentRun == FreeClusters &&
FatEntry != FAT_CLUSTER_AVAILABLE) {
ClustersThisRun = FatIndex - StartIndexOfThisRun;
if (FreeClusterCount) {
*FreeClusterCount += ClustersThisRun;
CurrentWindow->ClustersFree += ClustersThisRun;
}
if (BitMap) {
RtlClearBits( BitMap,
StartIndexOfThisRun - StartIndex,
ClustersThisRun );
}
CurrentRun = AllocatedClusters;
StartIndexOfThisRun = FatIndex;
}
//
// Are we switching from an allocated run to a free run?
//
if (CurrentRun == AllocatedClusters &&
FatEntry == FAT_CLUSTER_AVAILABLE) {
ClustersThisRun = FatIndex - StartIndexOfThisRun;
if (BitMap) {
RtlSetBits( BitMap,
StartIndexOfThisRun - StartIndex,
ClustersThisRun );
}
CurrentRun = FreeClusters;
StartIndexOfThisRun = FatIndex;
}
//
// If the entry is marked bad, add it to the bad block MCB
//
if ((SetupWindows || (Vcb->NumberOfWindows == 1)) &&
(FatInterpretClusterType( Vcb, FatEntry ) == FatClusterBad)) {
//
// This cluster is marked bad.
// Add it to the BadBlockMcb.
//
Lbo = FatGetLboFromIndex( Vcb, FatIndex );
FatAddMcbEntry( Vcb, &Vcb->BadBlockMcb, BadClusterVbo, Lbo, ClusterSize );
BadClusterVbo += ClusterSize;
}
}
//
// If we finished the scan, then we know about all the possible bad clusters.
//
SetFlag( Vcb->VcbState, VCB_STATE_FLAG_BAD_BLOCKS_POPULATED);
//
// Now we have to record the final run we encountered
//
ClustersThisRun = FatIndex - StartIndexOfThisRun;
if (CurrentRun == FreeClusters) {
if (FreeClusterCount) {
*FreeClusterCount += ClustersThisRun;
CurrentWindow->ClustersFree += ClustersThisRun;
}
if (BitMap) {
RtlClearBits( BitMap,
StartIndexOfThisRun - StartIndex,
ClustersThisRun );
}
} else {
if (BitMap) {
RtlSetBits( BitMap,
StartIndexOfThisRun - StartIndex,
ClustersThisRun );
}
}
//
// And finish the last window if we are in setup.
//
if (SetupWindows) {
CurrentWindow->LastCluster = FatIndex - 1;
}
//
// Now switch the active window if required. We've succesfully gotten everything
// nailed down.
//
// If we were tracking the free cluster count, this means we should update the
// window. This is the case of FAT12/16 initialization.
//
if (SwitchToWindow) {
if (Vcb->FreeClusterBitMap.Buffer) {
ExFreePool( Vcb->FreeClusterBitMap.Buffer );
}
RtlInitializeBitMap( &Vcb->FreeClusterBitMap,
NewBitMapBuffer,
EndIndex - StartIndex + 1 );
NewBitMapBuffer = NULL;
Vcb->CurrentWindow = SwitchToWindow;
Vcb->ClusterHint = (ULONG)-1;
if (FreeClusterCount) {
NT_ASSERT( !SetupWindows );
Vcb->CurrentWindow->ClustersFree = *FreeClusterCount;
}
}
//
// Make sure plausible things occured ...
//
if (!SetupWindows && BitMapBuffer == NULL) {
ASSERT_CURRENT_WINDOW_GOOD( Vcb );
}
NT_ASSERT(Vcb->AllocationSupport.NumberOfFreeClusters <= Vcb->AllocationSupport.NumberOfClusters);
} _SEH2_FINALLY {
//
// Unpin the last bcb and drop the temporary bitmap buffer if it exists.
//
FatUnpinBcb( IrpContext, Bcb);
if (NewBitMapBuffer) {
ExFreePool( NewBitMapBuffer );
}
} _SEH2_END;
}