mirror of
https://github.com/reactos/reactos.git
synced 2024-12-29 10:35:28 +00:00
0daa5547d9
This implies that a sample for W10. It has been backported to NT5.2; not sure how it would work on a W2K3 (feel free to test!)
5295 lines
158 KiB
C
5295 lines
158 KiB
C
/*++
|
||
|
||
Copyright (c) 1990-2000 Microsoft Corporation
|
||
|
||
Module Name:
|
||
|
||
AllocSup.c
|
||
|
||
Abstract:
|
||
|
||
This module implements the Allocation support routines for Fat.
|
||
|
||
|
||
--*/
|
||
|
||
#include "fatprocs.h"
|
||
|
||
//
|
||
// The Bug check file id for this module
|
||
//
|
||
|
||
#define BugCheckFileId (FAT_BUG_CHECK_ALLOCSUP)
|
||
|
||
//
|
||
// Local debug trace level
|
||
//
|
||
|
||
#define Dbg (DEBUG_TRACE_ALLOCSUP)
|
||
|
||
#define FatMin(a, b) ((a) < (b) ? (a) : (b))
|
||
|
||
//
|
||
// Define prefetch page count for the FAT
|
||
//
|
||
|
||
#define FAT_PREFETCH_PAGE_COUNT 0x100
|
||
|
||
//
|
||
// Local support routine prototypes
|
||
//
|
||
|
||
VOID
|
||
FatLookupFatEntry(
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG FatIndex,
|
||
IN OUT PULONG FatEntry,
|
||
IN OUT PFAT_ENUMERATION_CONTEXT Context
|
||
);
|
||
|
||
VOID
|
||
FatSetFatRun(
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG StartingFatIndex,
|
||
IN ULONG ClusterCount,
|
||
IN BOOLEAN ChainTogether
|
||
);
|
||
|
||
UCHAR
|
||
FatLogOf(
|
||
IN ULONG Value
|
||
);
|
||
|
||
//
|
||
// Note that the KdPrint below will ONLY fire when the assert does. Leave it
|
||
// alone.
|
||
//
|
||
|
||
#if DBG
|
||
#define ASSERT_CURRENT_WINDOW_GOOD(VCB) { \
|
||
ULONG FreeClusterBitMapClear; \
|
||
NT_ASSERT( (VCB)->FreeClusterBitMap.Buffer != NULL ); \
|
||
FreeClusterBitMapClear = RtlNumberOfClearBits(&(VCB)->FreeClusterBitMap); \
|
||
if ((VCB)->CurrentWindow->ClustersFree != FreeClusterBitMapClear) { \
|
||
KdPrint(("FAT: ClustersFree %x h != FreeClusterBitMapClear %x h\n", \
|
||
(VCB)->CurrentWindow->ClustersFree, \
|
||
FreeClusterBitMapClear)); \
|
||
} \
|
||
NT_ASSERT( (VCB)->CurrentWindow->ClustersFree == FreeClusterBitMapClear ); \
|
||
}
|
||
#else
|
||
#define ASSERT_CURRENT_WINDOW_GOOD(VCB)
|
||
#endif
|
||
|
||
//
|
||
// The following macros provide a convenient way of hiding the details
|
||
// of bitmap allocation schemes.
|
||
//
|
||
|
||
|
||
//
|
||
// VOID
|
||
// FatLockFreeClusterBitMap (
|
||
// IN PVCB Vcb
|
||
// );
|
||
//
|
||
|
||
#define FatLockFreeClusterBitMap(VCB) { \
|
||
NT_ASSERT(KeAreApcsDisabled()); \
|
||
ExAcquireFastMutexUnsafe( &(VCB)->FreeClusterBitMapMutex ); \
|
||
ASSERT_CURRENT_WINDOW_GOOD(VCB) \
|
||
}
|
||
|
||
//
|
||
// VOID
|
||
// FatUnlockFreeClusterBitMap (
|
||
// IN PVCB Vcb
|
||
// );
|
||
//
|
||
|
||
#define FatUnlockFreeClusterBitMap(VCB) { \
|
||
ASSERT_CURRENT_WINDOW_GOOD(VCB) \
|
||
NT_ASSERT(KeAreApcsDisabled()); \
|
||
ExReleaseFastMutexUnsafe( &(VCB)->FreeClusterBitMapMutex ); \
|
||
}
|
||
|
||
//
|
||
// BOOLEAN
|
||
// FatIsClusterFree (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG FatIndex
|
||
// );
|
||
//
|
||
|
||
#define FatIsClusterFree(IRPCONTEXT,VCB,FAT_INDEX) \
|
||
(RtlCheckBit(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2) == 0)
|
||
|
||
//
|
||
// VOID
|
||
// FatFreeClusters (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG FatIndex,
|
||
// IN ULONG ClusterCount
|
||
// );
|
||
//
|
||
|
||
#define FatFreeClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
|
||
if ((CLUSTER_COUNT) == 1) { \
|
||
FatSetFatEntry((IRPCONTEXT),(VCB),(FAT_INDEX),FAT_CLUSTER_AVAILABLE); \
|
||
} else { \
|
||
FatSetFatRun((IRPCONTEXT),(VCB),(FAT_INDEX),(CLUSTER_COUNT),FALSE); \
|
||
} \
|
||
}
|
||
|
||
//
|
||
// VOID
|
||
// FatAllocateClusters (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG FatIndex,
|
||
// IN ULONG ClusterCount
|
||
// );
|
||
//
|
||
|
||
#define FatAllocateClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
|
||
if ((CLUSTER_COUNT) == 1) { \
|
||
FatSetFatEntry((IRPCONTEXT),(VCB),(FAT_INDEX),FAT_CLUSTER_LAST); \
|
||
} else { \
|
||
FatSetFatRun((IRPCONTEXT),(VCB),(FAT_INDEX),(CLUSTER_COUNT),TRUE); \
|
||
} \
|
||
}
|
||
|
||
//
|
||
// VOID
|
||
// FatUnreserveClusters (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG FatIndex,
|
||
// IN ULONG ClusterCount
|
||
// );
|
||
//
|
||
|
||
#define FatUnreserveClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
|
||
NT_ASSERT( (FAT_INDEX) + (CLUSTER_COUNT) - 2 <= (VCB)->FreeClusterBitMap.SizeOfBitMap );\
|
||
NT_ASSERT( (FAT_INDEX) >= 2); \
|
||
RtlClearBits(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2,(CLUSTER_COUNT)); \
|
||
if ((FAT_INDEX) < (VCB)->ClusterHint) { \
|
||
(VCB)->ClusterHint = (FAT_INDEX); \
|
||
} \
|
||
}
|
||
|
||
//
|
||
// VOID
|
||
// FatReserveClusters (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG FatIndex,
|
||
// IN ULONG ClusterCount
|
||
// );
|
||
//
|
||
// Handle wrapping the hint back to the front.
|
||
//
|
||
|
||
#define FatReserveClusters(IRPCONTEXT,VCB,FAT_INDEX,CLUSTER_COUNT) { \
|
||
ULONG _AfterRun = (FAT_INDEX) + (CLUSTER_COUNT); \
|
||
NT_ASSERT( (FAT_INDEX) + (CLUSTER_COUNT) - 2 <= (VCB)->FreeClusterBitMap.SizeOfBitMap );\
|
||
NT_ASSERT( (FAT_INDEX) >= 2); \
|
||
RtlSetBits(&(VCB)->FreeClusterBitMap,(FAT_INDEX)-2,(CLUSTER_COUNT)); \
|
||
\
|
||
if (_AfterRun - 2 >= (VCB)->FreeClusterBitMap.SizeOfBitMap) { \
|
||
_AfterRun = 2; \
|
||
} \
|
||
if (RtlCheckBit(&(VCB)->FreeClusterBitMap, _AfterRun - 2)) { \
|
||
(VCB)->ClusterHint = RtlFindClearBits( &(VCB)->FreeClusterBitMap, 1, _AfterRun - 2) + 2; \
|
||
if (1 == (VCB)->ClusterHint) { \
|
||
(VCB)->ClusterHint = 2; \
|
||
} \
|
||
} \
|
||
else { \
|
||
(VCB)->ClusterHint = _AfterRun; \
|
||
} \
|
||
}
|
||
|
||
//
|
||
// ULONG
|
||
// FatFindFreeClusterRun (
|
||
// IN PIRP_CONTEXT IrpContext,
|
||
// IN PVCB Vcb,
|
||
// IN ULONG ClusterCount,
|
||
// IN ULONG AlternateClusterHint
|
||
// );
|
||
//
|
||
// Do a special check if only one cluster is desired.
|
||
//
|
||
|
||
#define FatFindFreeClusterRun(IRPCONTEXT,VCB,CLUSTER_COUNT,CLUSTER_HINT) ( \
|
||
(CLUSTER_COUNT == 1) && \
|
||
FatIsClusterFree((IRPCONTEXT), (VCB), (CLUSTER_HINT)) ? \
|
||
(CLUSTER_HINT) : \
|
||
RtlFindClearBits( &(VCB)->FreeClusterBitMap, \
|
||
(CLUSTER_COUNT), \
|
||
(CLUSTER_HINT) - 2) + 2 \
|
||
)
|
||
|
||
//
|
||
// FAT32: Define the maximum size of the FreeClusterBitMap to be the
|
||
// maximum size of a FAT16 FAT. If there are more clusters on the
|
||
// volume than can be represented by this many bytes of bitmap, the
|
||
// FAT will be split into "buckets", each of which does fit.
|
||
//
|
||
// Note this count is in clusters/bits of bitmap.
|
||
//
|
||
|
||
#define MAX_CLUSTER_BITMAP_SIZE (1 << 16)
|
||
|
||
//
|
||
// Calculate the window a given cluster number is in.
|
||
//
|
||
|
||
#define FatWindowOfCluster(C) (((C) - 2) / MAX_CLUSTER_BITMAP_SIZE)
|
||
|
||
#ifdef ALLOC_PRAGMA
|
||
#pragma alloc_text(PAGE, FatAddFileAllocation)
|
||
#pragma alloc_text(PAGE, FatAllocateDiskSpace)
|
||
#pragma alloc_text(PAGE, FatDeallocateDiskSpace)
|
||
#pragma alloc_text(PAGE, FatExamineFatEntries)
|
||
#pragma alloc_text(PAGE, FatInterpretClusterType)
|
||
#pragma alloc_text(PAGE, FatLogOf)
|
||
#pragma alloc_text(PAGE, FatLookupFatEntry)
|
||
#pragma alloc_text(PAGE, FatLookupFileAllocation)
|
||
#pragma alloc_text(PAGE, FatLookupFileAllocationSize)
|
||
#pragma alloc_text(PAGE, FatMergeAllocation)
|
||
#pragma alloc_text(PAGE, FatSetFatEntry)
|
||
#pragma alloc_text(PAGE, FatSetFatRun)
|
||
#pragma alloc_text(PAGE, FatSetupAllocationSupport)
|
||
#pragma alloc_text(PAGE, FatSplitAllocation)
|
||
#pragma alloc_text(PAGE, FatTearDownAllocationSupport)
|
||
#pragma alloc_text(PAGE, FatTruncateFileAllocation)
|
||
#endif
|
||
|
||
|
||
INLINE
|
||
ULONG
|
||
FatSelectBestWindow(
|
||
IN PVCB Vcb
|
||
)
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
Choose a window to allocate clusters from. Order of preference is:
|
||
|
||
1. First window with >50% free clusters
|
||
2. First empty window
|
||
3. Window with greatest number of free clusters.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb for the volume
|
||
|
||
Return Value:
|
||
|
||
'Best window' number (index into Vcb->Windows[])
|
||
|
||
--*/
|
||
{
|
||
ULONG i, Fave = 0;
|
||
ULONG MaxFree = 0;
|
||
ULONG FirstEmpty = (ULONG)-1;
|
||
ULONG ClustersPerWindow = MAX_CLUSTER_BITMAP_SIZE;
|
||
|
||
NT_ASSERT( 1 != Vcb->NumberOfWindows);
|
||
|
||
for (i = 0; i < Vcb->NumberOfWindows; i++) {
|
||
|
||
if (Vcb->Windows[i].ClustersFree == ClustersPerWindow) {
|
||
|
||
if (-1 == FirstEmpty) {
|
||
|
||
//
|
||
// Keep note of the first empty window on the disc
|
||
//
|
||
|
||
FirstEmpty = i;
|
||
}
|
||
}
|
||
else if (Vcb->Windows[i].ClustersFree > MaxFree) {
|
||
|
||
//
|
||
// This window has the most free clusters, so far
|
||
//
|
||
|
||
MaxFree = Vcb->Windows[i].ClustersFree;
|
||
Fave = i;
|
||
|
||
//
|
||
// If this window has >50% free clusters, then we will take it,
|
||
// so don't bother considering more windows.
|
||
//
|
||
|
||
if (MaxFree >= (ClustersPerWindow >> 1)) {
|
||
|
||
break;
|
||
}
|
||
}
|
||
}
|
||
|
||
//
|
||
// If there were no windows with 50% or more freespace, then select the
|
||
// first empty window on the disc, if any - otherwise we'll just go with
|
||
// the one with the most free clusters.
|
||
//
|
||
|
||
if ((MaxFree < (ClustersPerWindow >> 1)) && (-1 != FirstEmpty)) {
|
||
|
||
Fave = FirstEmpty;
|
||
}
|
||
|
||
return Fave;
|
||
}
|
||
|
||
|
||
VOID
|
||
FatSetupAllocationSupport (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine fills in the Allocation Support structure in the Vcb.
|
||
Most entries are computed using fat.h macros supplied with data from
|
||
the Bios Parameter Block. The free cluster count, however, requires
|
||
going to the Fat and actually counting free sectors. At the same time
|
||
the free cluster bit map is initalized.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to fill in.
|
||
|
||
--*/
|
||
|
||
{
|
||
ULONG BitIndex;
|
||
ULONG ClustersDescribableByFat;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatSetupAllocationSupport\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
|
||
//
|
||
// Compute a number of fields for Vcb.AllocationSupport
|
||
//
|
||
|
||
Vcb->AllocationSupport.RootDirectoryLbo = FatRootDirectoryLbo( &Vcb->Bpb );
|
||
Vcb->AllocationSupport.RootDirectorySize = FatRootDirectorySize( &Vcb->Bpb );
|
||
|
||
Vcb->AllocationSupport.FileAreaLbo = FatFileAreaLbo( &Vcb->Bpb );
|
||
|
||
Vcb->AllocationSupport.NumberOfClusters = FatNumberOfClusters( &Vcb->Bpb );
|
||
|
||
Vcb->AllocationSupport.FatIndexBitSize = FatIndexBitSize( &Vcb->Bpb );
|
||
|
||
Vcb->AllocationSupport.LogOfBytesPerSector = FatLogOf(Vcb->Bpb.BytesPerSector);
|
||
Vcb->AllocationSupport.LogOfBytesPerCluster = FatLogOf(FatBytesPerCluster( &Vcb->Bpb ));
|
||
Vcb->AllocationSupport.NumberOfFreeClusters = 0;
|
||
|
||
|
||
//
|
||
// Deal with a bug in DOS 5 format, if the Fat is not big enough to
|
||
// describe all the clusters on the disk, reduce this number. We expect
|
||
// that fat32 volumes will not have this problem.
|
||
//
|
||
// Turns out this was not a good assumption. We have to do this always now.
|
||
//
|
||
|
||
ClustersDescribableByFat = ( ((FatIsFat32(Vcb)? Vcb->Bpb.LargeSectorsPerFat :
|
||
Vcb->Bpb.SectorsPerFat) *
|
||
Vcb->Bpb.BytesPerSector * 8)
|
||
/ FatIndexBitSize(&Vcb->Bpb) ) - 2;
|
||
|
||
if (Vcb->AllocationSupport.NumberOfClusters > ClustersDescribableByFat) {
|
||
|
||
Vcb->AllocationSupport.NumberOfClusters = ClustersDescribableByFat;
|
||
}
|
||
|
||
//
|
||
// Extend the virtual volume file to include the Fat
|
||
//
|
||
|
||
{
|
||
CC_FILE_SIZES FileSizes;
|
||
|
||
FileSizes.AllocationSize.QuadPart =
|
||
FileSizes.FileSize.QuadPart = (FatReservedBytes( &Vcb->Bpb ) +
|
||
FatBytesPerFat( &Vcb->Bpb ));
|
||
FileSizes.ValidDataLength = FatMaxLarge;
|
||
|
||
if ( Vcb->VirtualVolumeFile->PrivateCacheMap == NULL ) {
|
||
|
||
FatInitializeCacheMap( Vcb->VirtualVolumeFile,
|
||
&FileSizes,
|
||
TRUE,
|
||
&FatData.CacheManagerNoOpCallbacks,
|
||
Vcb );
|
||
|
||
} else {
|
||
|
||
CcSetFileSizes( Vcb->VirtualVolumeFile, &FileSizes );
|
||
}
|
||
}
|
||
|
||
_SEH2_TRY {
|
||
|
||
if (FatIsFat32(Vcb) &&
|
||
Vcb->AllocationSupport.NumberOfClusters > MAX_CLUSTER_BITMAP_SIZE) {
|
||
|
||
Vcb->NumberOfWindows = (Vcb->AllocationSupport.NumberOfClusters +
|
||
MAX_CLUSTER_BITMAP_SIZE - 1) /
|
||
MAX_CLUSTER_BITMAP_SIZE;
|
||
|
||
} else {
|
||
|
||
Vcb->NumberOfWindows = 1;
|
||
}
|
||
|
||
Vcb->Windows = FsRtlAllocatePoolWithTag( PagedPool,
|
||
Vcb->NumberOfWindows * sizeof(FAT_WINDOW),
|
||
TAG_FAT_WINDOW );
|
||
|
||
RtlInitializeBitMap( &Vcb->FreeClusterBitMap,
|
||
NULL,
|
||
0 );
|
||
|
||
//
|
||
// Chose a FAT window to begin operation in.
|
||
//
|
||
|
||
if (Vcb->NumberOfWindows > 1) {
|
||
|
||
//
|
||
// Read the fat and count up free clusters. We bias by the two reserved
|
||
// entries in the FAT.
|
||
//
|
||
|
||
FatExamineFatEntries( IrpContext, Vcb,
|
||
2,
|
||
Vcb->AllocationSupport.NumberOfClusters + 2 - 1,
|
||
TRUE,
|
||
NULL,
|
||
NULL);
|
||
|
||
|
||
//
|
||
// Pick a window to begin allocating from
|
||
//
|
||
|
||
Vcb->CurrentWindow = &Vcb->Windows[ FatSelectBestWindow( Vcb)];
|
||
|
||
} else {
|
||
|
||
Vcb->CurrentWindow = &Vcb->Windows[0];
|
||
|
||
//
|
||
// Carefully bias ourselves by the two reserved entries in the FAT.
|
||
//
|
||
|
||
Vcb->CurrentWindow->FirstCluster = 2;
|
||
Vcb->CurrentWindow->LastCluster = Vcb->AllocationSupport.NumberOfClusters + 2 - 1;
|
||
}
|
||
|
||
//
|
||
// Now transition to the FAT window we have chosen.
|
||
//
|
||
|
||
FatExamineFatEntries( IrpContext, Vcb,
|
||
0,
|
||
0,
|
||
FALSE,
|
||
Vcb->CurrentWindow,
|
||
NULL);
|
||
|
||
//
|
||
// Now set the ClusterHint to the first free bit in our favorite
|
||
// window (except the ClusterHint is off by two).
|
||
//
|
||
|
||
Vcb->ClusterHint =
|
||
(BitIndex = RtlFindClearBits( &Vcb->FreeClusterBitMap, 1, 0 )) != -1 ?
|
||
BitIndex + 2 : 2;
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatSetupAllocationSupport );
|
||
|
||
//
|
||
// If we hit an exception, back out.
|
||
//
|
||
|
||
if (_SEH2_AbnormalTermination()) {
|
||
|
||
FatTearDownAllocationSupport( IrpContext, Vcb );
|
||
}
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
VOID
|
||
FatTearDownAllocationSupport (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine prepares the volume for closing. Specifically, we must
|
||
release the free fat bit map buffer, and uninitialize the dirty fat
|
||
Mcb.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to fill in.
|
||
|
||
Return Value:
|
||
|
||
VOID
|
||
|
||
--*/
|
||
|
||
{
|
||
DebugTrace(+1, Dbg, "FatTearDownAllocationSupport\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
|
||
PAGED_CODE();
|
||
|
||
//
|
||
// If there are FAT buckets, free them.
|
||
//
|
||
|
||
if ( Vcb->Windows != NULL ) {
|
||
|
||
ExFreePool( Vcb->Windows );
|
||
Vcb->Windows = NULL;
|
||
}
|
||
|
||
//
|
||
// Free the memory associated with the free cluster bitmap.
|
||
//
|
||
|
||
if ( Vcb->FreeClusterBitMap.Buffer != NULL ) {
|
||
|
||
ExFreePool( Vcb->FreeClusterBitMap.Buffer );
|
||
|
||
//
|
||
// NULL this field as an flag.
|
||
//
|
||
|
||
Vcb->FreeClusterBitMap.Buffer = NULL;
|
||
}
|
||
|
||
//
|
||
// And remove all the runs in the dirty fat Mcb
|
||
//
|
||
|
||
FatRemoveMcbEntry( Vcb, &Vcb->DirtyFatMcb, 0, 0xFFFFFFFF );
|
||
|
||
DebugTrace(-1, Dbg, "FatTearDownAllocationSupport -> (VOID)\n", 0);
|
||
|
||
UNREFERENCED_PARAMETER( IrpContext );
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatLookupFileAllocation (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PFCB FcbOrDcb,
|
||
IN VBO Vbo,
|
||
OUT PLBO Lbo,
|
||
OUT PULONG ByteCount,
|
||
OUT PBOOLEAN Allocated,
|
||
OUT PBOOLEAN EndOnMax,
|
||
OUT PULONG Index
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine looks up the existing mapping of VBO to LBO for a
|
||
file/directory. The information it queries is either stored in the
|
||
mcb field of the fcb/dcb or it is stored on in the fat table and
|
||
needs to be retrieved and decoded, and updated in the mcb.
|
||
|
||
Arguments:
|
||
|
||
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being queried
|
||
|
||
Vbo - Supplies the VBO whose LBO we want returned
|
||
|
||
Lbo - Receives the LBO corresponding to the input Vbo if one exists
|
||
|
||
ByteCount - Receives the number of bytes within the run the run
|
||
that correpond between the input vbo and output lbo.
|
||
|
||
Allocated - Receives TRUE if the Vbo does have a corresponding Lbo
|
||
and FALSE otherwise.
|
||
|
||
EndOnMax - Receives TRUE if the run ends in the maximal FAT cluster,
|
||
which results in a fractional bytecount.
|
||
|
||
Index - Receives the Index of the run
|
||
|
||
--*/
|
||
|
||
{
|
||
VBO CurrentVbo;
|
||
LBO CurrentLbo;
|
||
LBO PriorLbo;
|
||
|
||
VBO FirstVboOfCurrentRun = 0;
|
||
LBO FirstLboOfCurrentRun;
|
||
|
||
BOOLEAN LastCluster;
|
||
ULONG Runs;
|
||
|
||
PVCB Vcb;
|
||
FAT_ENTRY FatEntry;
|
||
ULONG BytesPerCluster;
|
||
ULARGE_INTEGER BytesOnVolume;
|
||
|
||
FAT_ENUMERATION_CONTEXT Context;
|
||
|
||
PAGED_CODE();
|
||
|
||
Vcb = FcbOrDcb->Vcb;
|
||
|
||
|
||
DebugTrace(+1, Dbg, "FatLookupFileAllocation\n", 0);
|
||
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
|
||
DebugTrace( 0, Dbg, " Vbo = %8lx\n", Vbo);
|
||
DebugTrace( 0, Dbg, " pLbo = %8lx\n", Lbo);
|
||
DebugTrace( 0, Dbg, " pByteCount = %8lx\n", ByteCount);
|
||
DebugTrace( 0, Dbg, " pAllocated = %8lx\n", Allocated);
|
||
|
||
Context.Bcb = NULL;
|
||
|
||
*EndOnMax = FALSE;
|
||
|
||
//
|
||
// Check the trivial case that the mapping is already in our
|
||
// Mcb.
|
||
//
|
||
|
||
if ( FatLookupMcbEntry(Vcb, &FcbOrDcb->Mcb, Vbo, Lbo, ByteCount, Index) ) {
|
||
|
||
*Allocated = TRUE;
|
||
|
||
NT_ASSERT( *ByteCount != 0 );
|
||
|
||
//
|
||
// Detect the overflow case, trim and claim the condition.
|
||
//
|
||
|
||
if (Vbo + *ByteCount == 0) {
|
||
|
||
*EndOnMax = TRUE;
|
||
}
|
||
|
||
DebugTrace( 0, Dbg, "Found run in Mcb.\n", 0);
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
//
|
||
// Initialize the Vcb, the cluster size, LastCluster, and
|
||
// FirstLboOfCurrentRun (to be used as an indication of the first
|
||
// iteration through the following while loop).
|
||
//
|
||
|
||
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
|
||
|
||
BytesOnVolume.QuadPart = UInt32x32To64( Vcb->AllocationSupport.NumberOfClusters, BytesPerCluster );
|
||
|
||
LastCluster = FALSE;
|
||
FirstLboOfCurrentRun = 0;
|
||
|
||
//
|
||
// Discard the case that the request extends beyond the end of
|
||
// allocation. Note that if the allocation size if not known
|
||
// AllocationSize is set to 0xffffffff.
|
||
//
|
||
|
||
if ( Vbo >= FcbOrDcb->Header.AllocationSize.LowPart ) {
|
||
|
||
*Allocated = FALSE;
|
||
|
||
DebugTrace( 0, Dbg, "Vbo beyond end of file.\n", 0);
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
//
|
||
// The Vbo is beyond the last Mcb entry. So we adjust Current Vbo/Lbo
|
||
// and FatEntry to describe the beginning of the last entry in the Mcb.
|
||
// This is used as initialization for the following loop.
|
||
//
|
||
// If the Mcb was empty, we start at the beginning of the file with
|
||
// CurrentVbo set to 0 to indicate a new run.
|
||
//
|
||
|
||
if (FatLookupLastMcbEntry( Vcb, &FcbOrDcb->Mcb, &CurrentVbo, &CurrentLbo, &Runs )) {
|
||
|
||
DebugTrace( 0, Dbg, "Current Mcb size = %8lx.\n", CurrentVbo + 1);
|
||
|
||
CurrentVbo -= (BytesPerCluster - 1);
|
||
CurrentLbo -= (BytesPerCluster - 1);
|
||
|
||
//
|
||
// Convert an index to a count.
|
||
//
|
||
|
||
Runs += 1;
|
||
|
||
} else {
|
||
|
||
DebugTrace( 0, Dbg, "Mcb empty.\n", 0);
|
||
|
||
//
|
||
// Check for an FcbOrDcb that has no allocation
|
||
//
|
||
|
||
if (FcbOrDcb->FirstClusterOfFile == 0) {
|
||
|
||
*Allocated = FALSE;
|
||
|
||
DebugTrace( 0, Dbg, "File has no allocation.\n", 0);
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
|
||
} else {
|
||
|
||
CurrentVbo = 0;
|
||
CurrentLbo = FatGetLboFromIndex( Vcb, FcbOrDcb->FirstClusterOfFile );
|
||
FirstVboOfCurrentRun = CurrentVbo;
|
||
FirstLboOfCurrentRun = CurrentLbo;
|
||
|
||
Runs = 0;
|
||
|
||
DebugTrace( 0, Dbg, "First Lbo of file = %8lx\n", CurrentLbo);
|
||
}
|
||
}
|
||
|
||
//
|
||
// Now we know that we are looking up a valid Vbo, but it is
|
||
// not in the Mcb, which is a monotonically increasing list of
|
||
// Vbo's. Thus we have to go to the Fat, and update
|
||
// the Mcb as we go. We use a try-finally to unpin the page
|
||
// of fat hanging around. Also we mark *Allocated = FALSE, so that
|
||
// the caller wont try to use the data if we hit an exception.
|
||
//
|
||
|
||
*Allocated = FALSE;
|
||
|
||
_SEH2_TRY {
|
||
|
||
FatEntry = (FAT_ENTRY)FatGetIndexFromLbo( Vcb, CurrentLbo );
|
||
|
||
//
|
||
// ASSERT that CurrentVbo and CurrentLbo are now cluster alligned.
|
||
// The assumption here, is that only whole clusters of Vbos and Lbos
|
||
// are mapped in the Mcb.
|
||
//
|
||
|
||
NT_ASSERT( ((CurrentLbo - Vcb->AllocationSupport.FileAreaLbo)
|
||
% BytesPerCluster == 0) &&
|
||
(CurrentVbo % BytesPerCluster == 0) );
|
||
|
||
//
|
||
// Starting from the first Vbo after the last Mcb entry, scan through
|
||
// the Fat looking for our Vbo. We continue through the Fat until we
|
||
// hit a noncontiguity beyond the desired Vbo, or the last cluster.
|
||
//
|
||
|
||
while ( !LastCluster ) {
|
||
|
||
//
|
||
// Get the next fat entry, and update our Current variables.
|
||
//
|
||
|
||
FatLookupFatEntry( IrpContext, Vcb, FatEntry, (PULONG)&FatEntry, &Context );
|
||
|
||
PriorLbo = CurrentLbo;
|
||
CurrentLbo = FatGetLboFromIndex( Vcb, FatEntry );
|
||
CurrentVbo += BytesPerCluster;
|
||
|
||
switch ( FatInterpretClusterType( Vcb, FatEntry )) {
|
||
|
||
//
|
||
// Check for a break in the Fat allocation chain.
|
||
//
|
||
|
||
case FatClusterAvailable:
|
||
case FatClusterReserved:
|
||
case FatClusterBad:
|
||
|
||
DebugTrace( 0, Dbg, "Break in allocation chain, entry = %d\n", FatEntry);
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> Fat Corrupt. Raise Status.\n", 0);
|
||
|
||
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
|
||
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
|
||
break;
|
||
|
||
//
|
||
// If this is the last cluster, we must update the Mcb and
|
||
// exit the loop.
|
||
//
|
||
|
||
case FatClusterLast:
|
||
|
||
//
|
||
// Assert we know where the current run started. If the
|
||
// Mcb was empty when we were called, thenFirstLboOfCurrentRun
|
||
// was set to the start of the file. If the Mcb contained an
|
||
// entry, then FirstLboOfCurrentRun was set on the first
|
||
// iteration through the loop. Thus if FirstLboOfCurrentRun
|
||
// is 0, then there was an Mcb entry and we are on our first
|
||
// iteration, meaing that the last cluster in the Mcb was
|
||
// really the last allocated cluster, but we checked Vbo
|
||
// against AllocationSize, and found it OK, thus AllocationSize
|
||
// must be too large.
|
||
//
|
||
// Note that, when we finally arrive here, CurrentVbo is actually
|
||
// the first Vbo beyond the file allocation and CurrentLbo is
|
||
// meaningless.
|
||
//
|
||
|
||
DebugTrace( 0, Dbg, "Read last cluster of file.\n", 0);
|
||
|
||
//
|
||
// Detect the case of the maximal file. Note that this really isn't
|
||
// a proper Vbo - those are zero-based, and this is a one-based number.
|
||
// The maximal file, of 2^32 - 1 bytes, has a maximum byte offset of
|
||
// 2^32 - 2.
|
||
//
|
||
// Just so we don't get confused here.
|
||
//
|
||
|
||
if (CurrentVbo == 0) {
|
||
|
||
*EndOnMax = TRUE;
|
||
CurrentVbo -= 1;
|
||
}
|
||
|
||
LastCluster = TRUE;
|
||
|
||
if (FirstLboOfCurrentRun != 0 ) {
|
||
|
||
DebugTrace( 0, Dbg, "Adding a run to the Mcb.\n", 0);
|
||
DebugTrace( 0, Dbg, " Vbo = %08lx.\n", FirstVboOfCurrentRun);
|
||
DebugTrace( 0, Dbg, " Lbo = %08lx.\n", FirstLboOfCurrentRun);
|
||
DebugTrace( 0, Dbg, " Length = %08lx.\n", CurrentVbo - FirstVboOfCurrentRun);
|
||
|
||
(VOID)FatAddMcbEntry( Vcb,
|
||
&FcbOrDcb->Mcb,
|
||
FirstVboOfCurrentRun,
|
||
FirstLboOfCurrentRun,
|
||
CurrentVbo - FirstVboOfCurrentRun );
|
||
|
||
Runs += 1;
|
||
}
|
||
|
||
//
|
||
// Being at the end of allocation, make sure we have found
|
||
// the Vbo. If we haven't, seeing as we checked VBO
|
||
// against AllocationSize, the real disk allocation is less
|
||
// than that of AllocationSize. This comes about when the
|
||
// real allocation is not yet known, and AllocaitonSize
|
||
// contains MAXULONG.
|
||
//
|
||
// KLUDGE! - If we were called by FatLookupFileAllocationSize
|
||
// Vbo is set to MAXULONG - 1, and AllocationSize to the lookup
|
||
// hint. Thus we merrily go along looking for a match that isn't
|
||
// there, but in the meantime building an Mcb. If this is
|
||
// the case, fill in AllocationSize and return.
|
||
//
|
||
|
||
if ( Vbo == MAXULONG - 1 ) {
|
||
|
||
*Allocated = FALSE;
|
||
|
||
FcbOrDcb->Header.AllocationSize.QuadPart = CurrentVbo;
|
||
|
||
DebugTrace( 0, Dbg, "New file allocation size = %08lx.\n", CurrentVbo);
|
||
try_return ( NOTHING );
|
||
}
|
||
|
||
//
|
||
// We will lie ever so slightly if we really terminated on the
|
||
// maximal byte of a file. It is really allocated.
|
||
//
|
||
|
||
if (Vbo >= CurrentVbo && !*EndOnMax) {
|
||
|
||
*Allocated = FALSE;
|
||
try_return ( NOTHING );
|
||
}
|
||
|
||
break;
|
||
|
||
//
|
||
// This is a continuation in the chain. If the run has a
|
||
// discontiguity at this point, update the Mcb, and if we are beyond
|
||
// the desired Vbo, this is the end of the run, so set LastCluster
|
||
// and exit the loop.
|
||
//
|
||
|
||
case FatClusterNext:
|
||
|
||
//
|
||
// This is the loop check. The Vbo must not be bigger than the size of
|
||
// the volume, and the Vbo must not have a) wrapped and b) not been at the
|
||
// very last cluster in the chain, for the case of the maximal file.
|
||
//
|
||
|
||
if ( CurrentVbo == 0 ||
|
||
(BytesOnVolume.HighPart == 0 && CurrentVbo > BytesOnVolume.LowPart)) {
|
||
|
||
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
|
||
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
|
||
}
|
||
|
||
if ( PriorLbo + BytesPerCluster != CurrentLbo ) {
|
||
|
||
//
|
||
// Note that on the first time through the loop
|
||
// (FirstLboOfCurrentRun == 0), we don't add the
|
||
// run to the Mcb since it curresponds to the last
|
||
// run already stored in the Mcb.
|
||
//
|
||
|
||
if ( FirstLboOfCurrentRun != 0 ) {
|
||
|
||
DebugTrace( 0, Dbg, "Adding a run to the Mcb.\n", 0);
|
||
DebugTrace( 0, Dbg, " Vbo = %08lx.\n", FirstVboOfCurrentRun);
|
||
DebugTrace( 0, Dbg, " Lbo = %08lx.\n", FirstLboOfCurrentRun);
|
||
DebugTrace( 0, Dbg, " Length = %08lx.\n", CurrentVbo - FirstVboOfCurrentRun);
|
||
|
||
FatAddMcbEntry( Vcb,
|
||
&FcbOrDcb->Mcb,
|
||
FirstVboOfCurrentRun,
|
||
FirstLboOfCurrentRun,
|
||
CurrentVbo - FirstVboOfCurrentRun );
|
||
|
||
Runs += 1;
|
||
}
|
||
|
||
//
|
||
// Since we are at a run boundry, with CurrentLbo and
|
||
// CurrentVbo being the first cluster of the next run,
|
||
// we see if the run we just added encompases the desired
|
||
// Vbo, and if so exit. Otherwise we set up two new
|
||
// First*boOfCurrentRun, and continue.
|
||
//
|
||
|
||
if (CurrentVbo > Vbo) {
|
||
|
||
LastCluster = TRUE;
|
||
|
||
} else {
|
||
|
||
FirstVboOfCurrentRun = CurrentVbo;
|
||
FirstLboOfCurrentRun = CurrentLbo;
|
||
}
|
||
}
|
||
break;
|
||
|
||
default:
|
||
|
||
DebugTrace(0, Dbg, "Illegal Cluster Type.\n", FatEntry);
|
||
|
||
#ifdef _MSC_VER
|
||
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
|
||
#endif
|
||
FatBugCheck( 0, 0, 0 );
|
||
|
||
break;
|
||
|
||
} // switch()
|
||
} // while()
|
||
|
||
//
|
||
// Load up the return parameters.
|
||
//
|
||
// On exit from the loop, Vbo still contains the desired Vbo, and
|
||
// CurrentVbo is the first byte after the run that contained the
|
||
// desired Vbo.
|
||
//
|
||
|
||
*Allocated = TRUE;
|
||
|
||
*Lbo = FirstLboOfCurrentRun + (Vbo - FirstVboOfCurrentRun);
|
||
|
||
*ByteCount = CurrentVbo - Vbo;
|
||
|
||
if (ARGUMENT_PRESENT(Index)) {
|
||
|
||
//
|
||
// Note that Runs only needs to be accurate with respect to where we
|
||
// ended. Since partial-lookup cases will occur without exclusive
|
||
// synchronization, the Mcb itself may be much bigger by now.
|
||
//
|
||
|
||
*Index = Runs - 1;
|
||
}
|
||
|
||
try_exit: NOTHING;
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatLookupFileAllocation );
|
||
|
||
//
|
||
// We are done reading the Fat, so unpin the last page of fat
|
||
// that is hanging around
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Context.Bcb );
|
||
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocation -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatAddFileAllocation (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PFCB FcbOrDcb,
|
||
IN PFILE_OBJECT FileObject OPTIONAL,
|
||
IN ULONG DesiredAllocationSize
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine adds additional allocation to the specified file/directory.
|
||
Additional allocation is added by appending clusters to the file/directory.
|
||
|
||
If the file already has a sufficient allocation then this procedure
|
||
is effectively a noop.
|
||
|
||
Arguments:
|
||
|
||
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified.
|
||
This parameter must not specify the root dcb.
|
||
|
||
FileObject - If supplied inform the cache manager of the change.
|
||
|
||
DesiredAllocationSize - Supplies the minimum size, in bytes, that we want
|
||
allocated to the file/directory.
|
||
|
||
--*/
|
||
|
||
{
|
||
PVCB Vcb;
|
||
LARGE_MCB NewMcb = {0};
|
||
PLARGE_MCB McbToCleanup = NULL;
|
||
PDIRENT Dirent = NULL;
|
||
ULONG NewAllocation = 0;
|
||
PBCB Bcb = NULL;
|
||
BOOLEAN UnwindWeAllocatedDiskSpace = FALSE;
|
||
BOOLEAN UnwindAllocationSizeSet = FALSE;
|
||
BOOLEAN UnwindCacheManagerInformed = FALSE;
|
||
BOOLEAN UnwindWeInitializedMcb = FALSE;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatAddFileAllocation\n", 0);
|
||
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
|
||
DebugTrace( 0, Dbg, " DesiredAllocationSize = %8lx\n", DesiredAllocationSize);
|
||
|
||
Vcb = FcbOrDcb->Vcb;
|
||
|
||
//
|
||
// If we haven't yet set the correct AllocationSize, do so.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
|
||
|
||
FatLookupFileAllocationSize( IrpContext, FcbOrDcb );
|
||
}
|
||
|
||
//
|
||
// Check for the benign case that the desired allocation is already
|
||
// within the allocation size.
|
||
//
|
||
|
||
if (DesiredAllocationSize <= FcbOrDcb->Header.AllocationSize.LowPart) {
|
||
|
||
DebugTrace(0, Dbg, "Desired size within current allocation.\n", 0);
|
||
|
||
DebugTrace(-1, Dbg, "FatAddFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
DebugTrace( 0, Dbg, "InitialAllocation = %08lx.\n", FcbOrDcb->Header.AllocationSize.LowPart);
|
||
|
||
//
|
||
// Get a chunk of disk space that will fullfill our needs. If there
|
||
// was no initial allocation, start from the hint in the Vcb, otherwise
|
||
// try to allocate from the cluster after the initial allocation.
|
||
//
|
||
// If there was no initial allocation to the file, we can just use the
|
||
// Mcb in the FcbOrDcb, otherwise we have to use a new one, and merge
|
||
// it to the one in the FcbOrDcb.
|
||
//
|
||
|
||
_SEH2_TRY {
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
|
||
|
||
LBO FirstLboOfFile;
|
||
|
||
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
|
||
|
||
FatGetDirentFromFcbOrDcb( IrpContext,
|
||
FcbOrDcb,
|
||
FALSE,
|
||
&Dirent,
|
||
&Bcb );
|
||
//
|
||
// Set this dirty right now since this call can fail.
|
||
//
|
||
|
||
FatSetDirtyBcb( IrpContext, Bcb, Vcb, TRUE );
|
||
|
||
FatAllocateDiskSpace( IrpContext,
|
||
Vcb,
|
||
0,
|
||
&DesiredAllocationSize,
|
||
FALSE,
|
||
&FcbOrDcb->Mcb );
|
||
|
||
UnwindWeAllocatedDiskSpace = TRUE;
|
||
McbToCleanup = &FcbOrDcb->Mcb;
|
||
|
||
//
|
||
// We have to update the dirent and FcbOrDcb copies of
|
||
// FirstClusterOfFile since before it was 0
|
||
//
|
||
|
||
FatLookupMcbEntry( FcbOrDcb->Vcb,
|
||
&FcbOrDcb->Mcb,
|
||
0,
|
||
&FirstLboOfFile,
|
||
(PULONG)NULL,
|
||
NULL );
|
||
|
||
DebugTrace( 0, Dbg, "First Lbo of file will be %08lx.\n", FirstLboOfFile );
|
||
|
||
FcbOrDcb->FirstClusterOfFile = FatGetIndexFromLbo( Vcb, FirstLboOfFile );
|
||
|
||
Dirent->FirstClusterOfFile = (USHORT)FcbOrDcb->FirstClusterOfFile;
|
||
|
||
if ( FatIsFat32(Vcb) ) {
|
||
|
||
Dirent->FirstClusterOfFileHi = (USHORT)(FcbOrDcb->FirstClusterOfFile >> 16);
|
||
}
|
||
|
||
//
|
||
// Note the size of the allocation we need to tell the cache manager about.
|
||
//
|
||
|
||
NewAllocation = DesiredAllocationSize;
|
||
|
||
} else {
|
||
|
||
LBO LastAllocatedLbo;
|
||
VBO DontCare;
|
||
|
||
//
|
||
// Get the first cluster following the current allocation. It is possible
|
||
// the Mcb is empty (or short, etc.) so we need to be slightly careful
|
||
// about making sure we don't lie with the hint.
|
||
//
|
||
|
||
(void)FatLookupLastMcbEntry( FcbOrDcb->Vcb, &FcbOrDcb->Mcb, &DontCare, &LastAllocatedLbo, NULL );
|
||
|
||
//
|
||
// Try to get some disk space starting from there.
|
||
//
|
||
|
||
NewAllocation = DesiredAllocationSize - FcbOrDcb->Header.AllocationSize.LowPart;
|
||
|
||
FsRtlInitializeLargeMcb( &NewMcb, PagedPool );
|
||
UnwindWeInitializedMcb = TRUE;
|
||
McbToCleanup = &NewMcb;
|
||
|
||
FatAllocateDiskSpace( IrpContext,
|
||
Vcb,
|
||
(LastAllocatedLbo != ~0 ?
|
||
FatGetIndexFromLbo(Vcb,LastAllocatedLbo + 1) :
|
||
0),
|
||
&NewAllocation,
|
||
FALSE,
|
||
&NewMcb );
|
||
|
||
UnwindWeAllocatedDiskSpace = TRUE;
|
||
}
|
||
|
||
//
|
||
// Now that we increased the allocation of the file, mark it in the
|
||
// FcbOrDcb. Carefully prepare to handle an inability to grow the cache
|
||
// structures.
|
||
//
|
||
|
||
FcbOrDcb->Header.AllocationSize.LowPart += NewAllocation;
|
||
|
||
//
|
||
// Handle the maximal file case, where we may have just wrapped. Note
|
||
// that this must be the precise boundary case wrap, i.e. by one byte,
|
||
// so that the new allocation is actually one byte "less" as far as we're
|
||
// concerned. This is important for the extension case.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
|
||
|
||
NewAllocation -= 1;
|
||
FcbOrDcb->Header.AllocationSize.LowPart = 0xffffffff;
|
||
}
|
||
|
||
UnwindAllocationSizeSet = TRUE;
|
||
|
||
//
|
||
// Inform the cache manager to increase the section size
|
||
//
|
||
|
||
if ( ARGUMENT_PRESENT(FileObject) && CcIsFileCached(FileObject) ) {
|
||
|
||
CcSetFileSizes( FileObject,
|
||
(PCC_FILE_SIZES)&FcbOrDcb->Header.AllocationSize );
|
||
UnwindCacheManagerInformed = TRUE;
|
||
}
|
||
|
||
//
|
||
// In the extension case, we have held off actually gluing the new
|
||
// allocation onto the file. This simplifies exception cleanup since
|
||
// if it was already added and the section grow failed, we'd have to
|
||
// do extra work to unglue it. This way, we can assume that if we
|
||
// raise the only thing we need to do is deallocate the disk space.
|
||
//
|
||
// Merge the allocation now.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.LowPart != NewAllocation) {
|
||
|
||
//
|
||
// Tack the new Mcb onto the end of the FcbOrDcb one.
|
||
//
|
||
|
||
FatMergeAllocation( IrpContext,
|
||
Vcb,
|
||
&FcbOrDcb->Mcb,
|
||
&NewMcb );
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatAddFileAllocation );
|
||
|
||
//
|
||
// Give FlushFileBuffer/Cleanup a clue here, regardless of success/fail..
|
||
//
|
||
|
||
SetFlag(FcbOrDcb->FcbState, FCB_STATE_FLUSH_FAT);
|
||
|
||
//
|
||
// If we were dogged trying to complete this operation, we need to go
|
||
// back various things out.
|
||
//
|
||
|
||
if (_SEH2_AbnormalTermination()) {
|
||
|
||
//
|
||
// Pull off the allocation size we tried to add to this object if
|
||
// we failed to grow cache structures or Mcb structures.
|
||
//
|
||
|
||
if (UnwindAllocationSizeSet) {
|
||
|
||
FcbOrDcb->Header.AllocationSize.LowPart -= NewAllocation;
|
||
}
|
||
|
||
if (UnwindCacheManagerInformed) {
|
||
|
||
CcSetFileSizes( FileObject,
|
||
(PCC_FILE_SIZES)&FcbOrDcb->Header.AllocationSize );
|
||
}
|
||
|
||
//
|
||
// In the case of initial allocation, we used the Fcb's Mcb and have
|
||
// to clean that up as well as the FAT chain references.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.LowPart == 0) {
|
||
|
||
if (Dirent != NULL) {
|
||
|
||
FcbOrDcb->FirstClusterOfFile = 0;
|
||
Dirent->FirstClusterOfFile = 0;
|
||
|
||
if ( FatIsFat32(Vcb) ) {
|
||
|
||
Dirent->FirstClusterOfFileHi = 0;
|
||
}
|
||
}
|
||
}
|
||
|
||
//
|
||
// ... and drop the dirent Bcb if we got it. Do it now
|
||
// so we can afford to take the exception if we have to.
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Bcb );
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Note this can re-raise.
|
||
//
|
||
|
||
if ( UnwindWeAllocatedDiskSpace ) {
|
||
|
||
FatDeallocateDiskSpace( IrpContext, Vcb, McbToCleanup, FALSE );
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
//
|
||
// We always want to clean up the non-initial allocation temporary Mcb,
|
||
// otherwise we have the Fcb's Mcb and we just truncate it away.
|
||
//
|
||
|
||
if (UnwindWeInitializedMcb == TRUE) {
|
||
|
||
//
|
||
// Note that we already know a raise is in progress. No danger
|
||
// of encountering the normal case code below and doing this again.
|
||
//
|
||
|
||
FsRtlUninitializeLargeMcb( McbToCleanup );
|
||
|
||
} else {
|
||
|
||
if (McbToCleanup) {
|
||
|
||
FsRtlTruncateLargeMcb( McbToCleanup, 0 );
|
||
}
|
||
}
|
||
} _SEH2_END;
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatAddFileAllocation -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
//
|
||
// Non-exceptional cleanup we always want to do. In handling the re-raise possibilities
|
||
// during exceptions we had to make sure these two steps always happened there beforehand.
|
||
// So now we handle the usual case.
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Bcb );
|
||
|
||
if (UnwindWeInitializedMcb == TRUE) {
|
||
|
||
FsRtlUninitializeLargeMcb( &NewMcb );
|
||
}
|
||
}
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatTruncateFileAllocation (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PFCB FcbOrDcb,
|
||
IN ULONG DesiredAllocationSize
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine truncates the allocation to the specified file/directory.
|
||
|
||
If the file is already smaller than the indicated size then this procedure
|
||
is effectively a noop.
|
||
|
||
|
||
Arguments:
|
||
|
||
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified
|
||
This parameter must not specify the root dcb.
|
||
|
||
DesiredAllocationSize - Supplies the maximum size, in bytes, that we want
|
||
allocated to the file/directory. It is rounded
|
||
up to the nearest cluster.
|
||
|
||
Return Value:
|
||
|
||
VOID - TRUE if the operation completed and FALSE if it had to
|
||
block but could not.
|
||
|
||
--*/
|
||
|
||
{
|
||
PVCB Vcb;
|
||
PBCB Bcb = NULL;
|
||
LARGE_MCB RemainingMcb = {0};
|
||
ULONG BytesPerCluster;
|
||
PDIRENT Dirent = NULL;
|
||
BOOLEAN UpdatedDirent = FALSE;
|
||
|
||
ULONG UnwindInitialAllocationSize;
|
||
ULONG UnwindInitialFirstClusterOfFile;
|
||
BOOLEAN UnwindWeAllocatedMcb = FALSE;
|
||
|
||
PAGED_CODE();
|
||
|
||
Vcb = FcbOrDcb->Vcb;
|
||
|
||
DebugTrace(+1, Dbg, "FatTruncateFileAllocation\n", 0);
|
||
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
|
||
DebugTrace( 0, Dbg, " DesiredAllocationSize = %8lx\n", DesiredAllocationSize);
|
||
|
||
//
|
||
// If the Fcb isn't in good condition, we have no business whacking around on
|
||
// the disk after "its" clusters.
|
||
//
|
||
// Inspired by a Prefix complaint.
|
||
//
|
||
|
||
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
|
||
|
||
//
|
||
// If we haven't yet set the correct AllocationSize, do so.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
|
||
|
||
FatLookupFileAllocationSize( IrpContext, FcbOrDcb );
|
||
}
|
||
|
||
//
|
||
// Round up the Desired Allocation Size to the next cluster size
|
||
//
|
||
|
||
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
|
||
|
||
//
|
||
// Note if the desired allocation is zero, to distinguish this from
|
||
// the wrap case below.
|
||
//
|
||
|
||
if (DesiredAllocationSize != 0) {
|
||
|
||
DesiredAllocationSize = (DesiredAllocationSize + (BytesPerCluster - 1)) &
|
||
~(BytesPerCluster - 1);
|
||
//
|
||
// Check for the benign case that the file is already smaller than
|
||
// the desired truncation. Note that if it wraps, then a) it was
|
||
// specifying an offset in the maximally allocatable cluster and
|
||
// b) we're not asking to extend the file, either. So stop.
|
||
//
|
||
|
||
if (DesiredAllocationSize == 0 ||
|
||
DesiredAllocationSize >= FcbOrDcb->Header.AllocationSize.LowPart) {
|
||
|
||
DebugTrace(0, Dbg, "Desired size within current allocation.\n", 0);
|
||
|
||
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
}
|
||
|
||
//
|
||
// This is a no-op if the allocation size is already what we want.
|
||
//
|
||
|
||
if (DesiredAllocationSize == FcbOrDcb->Header.AllocationSize.LowPart) {
|
||
|
||
DebugTrace(0, Dbg, "Desired size equals current allocation.\n", 0);
|
||
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
UnwindInitialAllocationSize = FcbOrDcb->Header.AllocationSize.LowPart;
|
||
UnwindInitialFirstClusterOfFile = FcbOrDcb->FirstClusterOfFile;
|
||
|
||
//
|
||
// Update the FcbOrDcb allocation size. If it is now zero, we have the
|
||
// additional task of modifying the FcbOrDcb and Dirent copies of
|
||
// FirstClusterInFile.
|
||
//
|
||
// Note that we must pin the dirent before actually deallocating the
|
||
// disk space since, in unwind, it would not be possible to reallocate
|
||
// deallocated disk space as someone else may have reallocated it and
|
||
// may cause an exception when you try to get some more disk space.
|
||
// Thus FatDeallocateDiskSpace must be the final dangerous operation.
|
||
//
|
||
|
||
_SEH2_TRY {
|
||
|
||
FcbOrDcb->Header.AllocationSize.QuadPart = DesiredAllocationSize;
|
||
|
||
//
|
||
// Special case 0
|
||
//
|
||
|
||
if (DesiredAllocationSize == 0) {
|
||
|
||
//
|
||
// We have to update the dirent and FcbOrDcb copies of
|
||
// FirstClusterOfFile since before it was 0
|
||
//
|
||
|
||
NT_ASSERT( FcbOrDcb->FcbCondition == FcbGood );
|
||
|
||
FatGetDirentFromFcbOrDcb( IrpContext, FcbOrDcb, FALSE, &Dirent, &Bcb );
|
||
|
||
Dirent->FirstClusterOfFile = 0;
|
||
|
||
if (FatIsFat32(Vcb)) {
|
||
|
||
Dirent->FirstClusterOfFileHi = 0;
|
||
}
|
||
|
||
FcbOrDcb->FirstClusterOfFile = 0;
|
||
|
||
FatSetDirtyBcb( IrpContext, Bcb, Vcb, TRUE );
|
||
UpdatedDirent = TRUE;
|
||
|
||
FatDeallocateDiskSpace( IrpContext, Vcb, &FcbOrDcb->Mcb, ((FcbOrDcb->FcbState & FCB_STATE_ZERO_ON_DEALLOCATION) != 0));
|
||
|
||
FatRemoveMcbEntry( FcbOrDcb->Vcb, &FcbOrDcb->Mcb, 0, 0xFFFFFFFF );
|
||
|
||
} else {
|
||
|
||
//
|
||
// Split the existing allocation into two parts, one we will keep, and
|
||
// one we will deallocate.
|
||
//
|
||
|
||
FsRtlInitializeLargeMcb( &RemainingMcb, PagedPool );
|
||
UnwindWeAllocatedMcb = TRUE;
|
||
|
||
FatSplitAllocation( IrpContext,
|
||
Vcb,
|
||
&FcbOrDcb->Mcb,
|
||
DesiredAllocationSize,
|
||
&RemainingMcb );
|
||
|
||
FatDeallocateDiskSpace( IrpContext, Vcb, &RemainingMcb, ((FcbOrDcb->FcbState & FCB_STATE_ZERO_ON_DEALLOCATION) != 0) );
|
||
|
||
FsRtlUninitializeLargeMcb( &RemainingMcb );
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatTruncateFileAllocation );
|
||
|
||
//
|
||
// Is this really the right backout strategy? It would be nice if we could
|
||
// pretend the truncate worked if we knew that the file had gotten into
|
||
// a consistent state. Leaving dangled clusters is probably quite preferable.
|
||
//
|
||
|
||
if ( _SEH2_AbnormalTermination() ) {
|
||
|
||
FcbOrDcb->Header.AllocationSize.LowPart = UnwindInitialAllocationSize;
|
||
|
||
if ( (DesiredAllocationSize == 0) && (Dirent != NULL)) {
|
||
|
||
if (UpdatedDirent) {
|
||
|
||
//
|
||
// If the dirent has been updated ok and marked dirty, then we
|
||
// failed in deallocatediscspace, and don't know what state
|
||
// the on disc fat chain is in. So we throw away the mcb,
|
||
// and potentially loose a few clusters until the next
|
||
// chkdsk. The operation has succeeded, but the exception
|
||
// will still propogate. 5.1
|
||
//
|
||
|
||
FatRemoveMcbEntry( Vcb, &FcbOrDcb->Mcb, 0, 0xFFFFFFFF );
|
||
FcbOrDcb->Header.AllocationSize.QuadPart = 0;
|
||
}
|
||
else if (FcbOrDcb->FirstClusterOfFile == 0) {
|
||
|
||
Dirent->FirstClusterOfFile = (USHORT)UnwindInitialFirstClusterOfFile;
|
||
|
||
if ( FatIsFat32(Vcb) ) {
|
||
|
||
Dirent->FirstClusterOfFileHi =
|
||
(USHORT)(UnwindInitialFirstClusterOfFile >> 16);
|
||
}
|
||
|
||
FcbOrDcb->FirstClusterOfFile = UnwindInitialFirstClusterOfFile;
|
||
}
|
||
}
|
||
|
||
if ( UnwindWeAllocatedMcb ) {
|
||
|
||
FsRtlUninitializeLargeMcb( &RemainingMcb );
|
||
}
|
||
|
||
//
|
||
// Note that in the non zero truncation case, we will also
|
||
// leak clusters. However, apart from this, the in memory and on disc
|
||
// structures will agree.
|
||
}
|
||
|
||
FatUnpinBcb( IrpContext, Bcb );
|
||
|
||
//
|
||
// Give FlushFileBuffer/Cleanup a clue here, regardless of success/fail.
|
||
//
|
||
|
||
SetFlag(FcbOrDcb->FcbState, FCB_STATE_FLUSH_FAT);
|
||
|
||
DebugTrace(-1, Dbg, "FatTruncateFileAllocation -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatLookupFileAllocationSize (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PFCB FcbOrDcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine retrieves the current file allocatio size for the
|
||
specified file/directory.
|
||
|
||
Arguments:
|
||
|
||
FcbOrDcb - Supplies the Fcb/Dcb of the file/directory being modified
|
||
|
||
--*/
|
||
|
||
{
|
||
LBO Lbo;
|
||
ULONG ByteCount;
|
||
BOOLEAN DontCare;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatLookupAllocationSize\n", 0);
|
||
DebugTrace( 0, Dbg, " FcbOrDcb = %p\n", FcbOrDcb);
|
||
|
||
//
|
||
// We call FatLookupFileAllocation with Vbo of 0xffffffff - 1.
|
||
//
|
||
|
||
FatLookupFileAllocation( IrpContext,
|
||
FcbOrDcb,
|
||
MAXULONG - 1,
|
||
&Lbo,
|
||
&ByteCount,
|
||
&DontCare,
|
||
&DontCare,
|
||
NULL );
|
||
|
||
//
|
||
// FileSize was set at Fcb creation time from the contents of the directory entry,
|
||
// and we are only now looking up the real length of the allocation chain. If it
|
||
// cannot be contained, this is trash. Probably more where that came from.
|
||
//
|
||
|
||
if (FcbOrDcb->Header.FileSize.LowPart > FcbOrDcb->Header.AllocationSize.LowPart) {
|
||
|
||
FatPopUpFileCorrupt( IrpContext, FcbOrDcb );
|
||
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatLookupFileAllocationSize -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatAllocateDiskSpace (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG AbsoluteClusterHint,
|
||
IN PULONG ByteCount,
|
||
IN BOOLEAN ExactMatchRequired,
|
||
OUT PLARGE_MCB Mcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This procedure allocates additional disk space and builds an mcb
|
||
representing the newly allocated space. If the space cannot be
|
||
allocated then this procedure raises an appropriate status.
|
||
|
||
Searching starts from the hint index in the Vcb unless an alternative
|
||
non-zero hint is given in AlternateClusterHint. If we are using the
|
||
hint field in the Vcb, it is set to the cluster following our allocation
|
||
when we are done.
|
||
|
||
Disk space can only be allocated in cluster units so this procedure
|
||
will round up any byte count to the next cluster boundary.
|
||
|
||
Pictorially what is done is the following (where ! denotes the end of
|
||
the fat chain (i.e., FAT_CLUSTER_LAST)):
|
||
|
||
|
||
Mcb (empty)
|
||
|
||
becomes
|
||
|
||
Mcb |--a--|--b--|--c--!
|
||
|
||
^
|
||
ByteCount ----------+
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the VCB being modified
|
||
|
||
AbsoluteClusterHint - Supplies an alternate hint index to start the
|
||
search from. If this is zero we use, and update,
|
||
the Vcb hint field.
|
||
|
||
ByteCount - Supplies the number of bytes that we are requesting, and
|
||
receives the number of bytes that we got.
|
||
|
||
ExactMatchRequired - Caller should set this to TRUE if only the precise run requested
|
||
is acceptable.
|
||
|
||
Mcb - Receives the MCB describing the newly allocated disk space. The
|
||
caller passes in an initialized Mcb that is filled in by this procedure.
|
||
|
||
Return Value:
|
||
|
||
TRUE - Allocated ok
|
||
FALSE - Failed to allocate exactly as requested (=> ExactMatchRequired was TRUE)
|
||
|
||
--*/
|
||
|
||
{
|
||
UCHAR LogOfBytesPerCluster;
|
||
ULONG BytesPerCluster;
|
||
ULONG StartingCluster;
|
||
ULONG ClusterCount;
|
||
ULONG WindowRelativeHint;
|
||
#if DBG
|
||
ULONG PreviousClear = 0;
|
||
#endif
|
||
|
||
PFAT_WINDOW Window;
|
||
BOOLEAN Wait = FALSE;
|
||
BOOLEAN Result = TRUE;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatAllocateDiskSpace\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " *ByteCount = %8lx\n", *ByteCount);
|
||
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
|
||
DebugTrace( 0, Dbg, " Hint = %8lx\n", AbsoluteClusterHint);
|
||
|
||
NT_ASSERT((AbsoluteClusterHint <= Vcb->AllocationSupport.NumberOfClusters + 2) && (1 != AbsoluteClusterHint));
|
||
|
||
//
|
||
// Make sure byte count is not zero
|
||
//
|
||
|
||
if (*ByteCount == 0) {
|
||
|
||
DebugTrace(0, Dbg, "Nothing to allocate.\n", 0);
|
||
|
||
DebugTrace(-1, Dbg, "FatAllocateDiskSpace -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
//
|
||
// Compute the cluster count based on the byte count, rounding up
|
||
// to the next cluster if there is any remainder. Note that the
|
||
// pathalogical case BytesCount == 0 has been eliminated above.
|
||
//
|
||
|
||
LogOfBytesPerCluster = Vcb->AllocationSupport.LogOfBytesPerCluster;
|
||
BytesPerCluster = 1 << LogOfBytesPerCluster;
|
||
|
||
*ByteCount = (*ByteCount + (BytesPerCluster - 1))
|
||
& ~(BytesPerCluster - 1);
|
||
|
||
//
|
||
// If ByteCount is NOW zero, then we were asked for the maximal
|
||
// filesize (or at least for bytes in the last allocatable sector).
|
||
//
|
||
|
||
if (*ByteCount == 0) {
|
||
|
||
*ByteCount = 0xffffffff;
|
||
ClusterCount = 1 << (32 - LogOfBytesPerCluster);
|
||
|
||
} else {
|
||
|
||
ClusterCount = (*ByteCount >> LogOfBytesPerCluster);
|
||
}
|
||
|
||
//
|
||
// Analysis tools don't figure out that ClusterCount is not zero because
|
||
// of the ByteCount == 0 checks, so give them a hint.
|
||
//
|
||
_Analysis_assume_(ClusterCount > 0);
|
||
|
||
//
|
||
// Make sure there are enough free clusters to start with, and
|
||
// take them now so that nobody else takes them from us.
|
||
//
|
||
|
||
ExAcquireResourceSharedLite(&Vcb->ChangeBitMapResource, TRUE);
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
|
||
if (ClusterCount <= Vcb->AllocationSupport.NumberOfFreeClusters) {
|
||
|
||
Vcb->AllocationSupport.NumberOfFreeClusters -= ClusterCount;
|
||
|
||
} else {
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
|
||
DebugTrace(0, Dbg, "Disk Full. Raise Status.\n", 0);
|
||
FatRaiseStatus( IrpContext, STATUS_DISK_FULL );
|
||
}
|
||
|
||
//
|
||
// Did the caller supply a hint?
|
||
//
|
||
|
||
if ((0 != AbsoluteClusterHint) && (AbsoluteClusterHint < (Vcb->AllocationSupport.NumberOfClusters + 2))) {
|
||
|
||
if (Vcb->NumberOfWindows > 1) {
|
||
|
||
//
|
||
// If we're being called upon to allocate clusters outside the
|
||
// current window (which happens only via MoveFile), it's a problem.
|
||
// We address this by changing the current window to be the one which
|
||
// contains the alternate cluster hint. Note that if the user's
|
||
// request would cross a window boundary, he doesn't really get what
|
||
// he wanted.
|
||
//
|
||
|
||
if (AbsoluteClusterHint < Vcb->CurrentWindow->FirstCluster ||
|
||
AbsoluteClusterHint > Vcb->CurrentWindow->LastCluster) {
|
||
|
||
ULONG BucketNum = FatWindowOfCluster( AbsoluteClusterHint );
|
||
|
||
NT_ASSERT( BucketNum < Vcb->NumberOfWindows);
|
||
|
||
//
|
||
// Drop our shared lock on the ChangeBitMapResource, and pick it up again
|
||
// exclusive in preparation for making the window swap.
|
||
//
|
||
|
||
FatUnlockFreeClusterBitMap(Vcb);
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
ExAcquireResourceExclusiveLite(&Vcb->ChangeBitMapResource, TRUE);
|
||
FatLockFreeClusterBitMap(Vcb);
|
||
|
||
Window = &Vcb->Windows[BucketNum];
|
||
|
||
//
|
||
// Again, test the current window against the one we want - some other
|
||
// thread could have sneaked in behind our backs and kindly set it to the one
|
||
// we need, when we dropped and reacquired the ChangeBitMapResource above.
|
||
//
|
||
|
||
if (Window != Vcb->CurrentWindow) {
|
||
|
||
_SEH2_TRY {
|
||
|
||
Wait = BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
SetFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
|
||
//
|
||
// Change to the new window (update Vcb->CurrentWindow) and scan it
|
||
// to build up a freespace bitmap etc.
|
||
//
|
||
|
||
FatExamineFatEntries( IrpContext, Vcb,
|
||
0,
|
||
0,
|
||
FALSE,
|
||
Window,
|
||
NULL);
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
if (!Wait) {
|
||
|
||
ClearFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
}
|
||
|
||
if (_SEH2_AbnormalTermination()) {
|
||
|
||
//
|
||
// We will have raised as a result of failing to pick up the
|
||
// chunk of the FAT for this window move. Release our resources
|
||
// and return the cluster count to the volume.
|
||
//
|
||
|
||
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
}
|
||
} _SEH2_END;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Make the hint cluster number relative to the base of the current window...
|
||
//
|
||
// Currentwindow->Firstcluster is baised by +2 already, so we will lose the
|
||
// bias already in AbsoluteClusterHint. Put it back....
|
||
//
|
||
|
||
WindowRelativeHint = AbsoluteClusterHint - Vcb->CurrentWindow->FirstCluster + 2;
|
||
}
|
||
else {
|
||
|
||
//
|
||
// Only one 'window', ie fat16/12. No modification necessary.
|
||
//
|
||
|
||
WindowRelativeHint = AbsoluteClusterHint;
|
||
}
|
||
}
|
||
else {
|
||
|
||
//
|
||
// Either no hint supplied, or it was out of range, so grab one from the Vcb
|
||
//
|
||
// NOTE: Clusterhint in the Vcb is not guaranteed to be set (may be -1)
|
||
//
|
||
|
||
WindowRelativeHint = Vcb->ClusterHint;
|
||
AbsoluteClusterHint = 0;
|
||
|
||
//
|
||
// Vcb hint may not have been initialized yet. Force to valid cluster.
|
||
//
|
||
|
||
if (-1 == WindowRelativeHint) {
|
||
|
||
WindowRelativeHint = 2;
|
||
}
|
||
}
|
||
|
||
NT_ASSERT((WindowRelativeHint >= 2) && (WindowRelativeHint < Vcb->FreeClusterBitMap.SizeOfBitMap + 2));
|
||
|
||
//
|
||
// Keep track of the window we're allocating from, so we can clean
|
||
// up correctly if the current window changes after we unlock the
|
||
// bitmap.
|
||
//
|
||
|
||
Window = Vcb->CurrentWindow;
|
||
|
||
//
|
||
// Try to find a run of free clusters large enough for us.
|
||
//
|
||
|
||
StartingCluster = FatFindFreeClusterRun( IrpContext,
|
||
Vcb,
|
||
ClusterCount,
|
||
WindowRelativeHint );
|
||
//
|
||
// If the above call was successful, we can just update the fat
|
||
// and Mcb and exit. Otherwise we have to look for smaller free
|
||
// runs.
|
||
//
|
||
// This test is a bit funky. Note that the error return from
|
||
// RtlFindClearBits is -1, and adding two to that is 1.
|
||
//
|
||
|
||
if ((StartingCluster != 1) &&
|
||
((0 == AbsoluteClusterHint) || (StartingCluster == WindowRelativeHint))
|
||
) {
|
||
|
||
#if DBG
|
||
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
|
||
#endif // DBG
|
||
|
||
//
|
||
// Take the clusters we found, and unlock the bit map.
|
||
//
|
||
|
||
FatReserveClusters(IrpContext, Vcb, StartingCluster, ClusterCount);
|
||
|
||
Window->ClustersFree -= ClusterCount;
|
||
|
||
StartingCluster += Window->FirstCluster;
|
||
StartingCluster -= 2;
|
||
|
||
NT_ASSERT( PreviousClear - ClusterCount == Window->ClustersFree );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
|
||
//
|
||
// Note that this call will never fail since there is always
|
||
// room for one entry in an empty Mcb.
|
||
//
|
||
|
||
FatAddMcbEntry( Vcb, Mcb,
|
||
0,
|
||
FatGetLboFromIndex( Vcb, StartingCluster ),
|
||
*ByteCount);
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Update the fat.
|
||
//
|
||
|
||
FatAllocateClusters(IrpContext, Vcb,
|
||
StartingCluster,
|
||
ClusterCount);
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatAllocateDiskSpace );
|
||
|
||
//
|
||
// If the allocate clusters failed, remove the run from the Mcb,
|
||
// unreserve the clusters, and reset the free cluster count.
|
||
//
|
||
|
||
if (_SEH2_AbnormalTermination()) {
|
||
|
||
FatRemoveMcbEntry( Vcb, Mcb, 0, *ByteCount );
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
|
||
// Only clear bits if the bitmap window is the same.
|
||
|
||
if (Window == Vcb->CurrentWindow) {
|
||
|
||
// Both values (startingcluster and window->firstcluster) are
|
||
// already biased by 2, so will cancel, so we need to add in the 2 again.
|
||
|
||
FatUnreserveClusters( IrpContext, Vcb,
|
||
StartingCluster - Window->FirstCluster + 2,
|
||
ClusterCount );
|
||
}
|
||
|
||
Window->ClustersFree += ClusterCount;
|
||
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
}
|
||
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
} _SEH2_END;
|
||
|
||
} else {
|
||
|
||
//
|
||
// Note that Index is a zero-based window-relative number. When appropriate
|
||
// it'll get converted into a true cluster number and put in Cluster, which
|
||
// will be a volume relative true cluster number.
|
||
//
|
||
|
||
ULONG Index = 0;
|
||
ULONG Cluster = 0;
|
||
ULONG CurrentVbo = 0;
|
||
ULONG PriorLastCluster = 0;
|
||
ULONG BytesFound = 0;
|
||
|
||
ULONG ClustersFound = 0;
|
||
ULONG ClustersRemaining = 0;
|
||
|
||
BOOLEAN LockedBitMap = FALSE;
|
||
BOOLEAN SelectNextContigWindow = FALSE;
|
||
|
||
//
|
||
// Drop our shared lock on the ChangeBitMapResource, and pick it up again
|
||
// exclusive in preparation for making a window swap.
|
||
//
|
||
|
||
FatUnlockFreeClusterBitMap(Vcb);
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
ExAcquireResourceExclusiveLite(&Vcb->ChangeBitMapResource, TRUE);
|
||
FatLockFreeClusterBitMap(Vcb);
|
||
LockedBitMap = TRUE;
|
||
|
||
_SEH2_TRY {
|
||
|
||
if ( ExactMatchRequired && (1 == Vcb->NumberOfWindows)) {
|
||
|
||
//
|
||
// Give up right now, there are no more windows to search! RtlFindClearBits
|
||
// searchs the whole bitmap, so we would have found any contiguous run
|
||
// large enough.
|
||
//
|
||
|
||
try_leave( Result = FALSE);
|
||
}
|
||
|
||
//
|
||
// While the request is still incomplete, look for the largest
|
||
// run of free clusters, mark them taken, allocate the run in
|
||
// the Mcb and Fat, and if this isn't the first time through
|
||
// the loop link it to prior run on the fat. The Mcb will
|
||
// coalesce automatically.
|
||
//
|
||
|
||
ClustersRemaining = ClusterCount;
|
||
CurrentVbo = 0;
|
||
PriorLastCluster = 0;
|
||
|
||
while (ClustersRemaining != 0) {
|
||
|
||
//
|
||
// If we just entered the loop, the bit map is already locked
|
||
//
|
||
|
||
if ( !LockedBitMap ) {
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
LockedBitMap = TRUE;
|
||
}
|
||
|
||
//
|
||
// Find the largest run of free clusters. If the run is
|
||
// bigger than we need, only use what we need. Note that
|
||
// this will then be the last while() iteration.
|
||
//
|
||
|
||
// 12/3/95: need to bias bitmap by 2 bits for the defrag
|
||
// hooks and the below macro became impossible to do without in-line
|
||
// procedures.
|
||
//
|
||
// ClustersFound = FatLongestFreeClusterRun( IrpContext, Vcb, &Index );
|
||
|
||
ClustersFound = 0;
|
||
|
||
if (!SelectNextContigWindow) {
|
||
|
||
if ( 0 != WindowRelativeHint) {
|
||
|
||
ULONG Desired = Vcb->FreeClusterBitMap.SizeOfBitMap - (WindowRelativeHint - 2);
|
||
|
||
//
|
||
// We will try to allocate contiguously. Try from the current hint the to
|
||
// end of current window. Don't try for more than we actually need.
|
||
//
|
||
|
||
if (Desired > ClustersRemaining) {
|
||
|
||
Desired = ClustersRemaining;
|
||
}
|
||
|
||
if (RtlAreBitsClear( &Vcb->FreeClusterBitMap,
|
||
WindowRelativeHint - 2,
|
||
Desired))
|
||
{
|
||
//
|
||
// Clusters from hint->...windowend are free. Take them.
|
||
//
|
||
|
||
Index = WindowRelativeHint - 2;
|
||
ClustersFound = Desired;
|
||
|
||
if (FatIsFat32(Vcb)) {
|
||
|
||
//
|
||
// We're now up against the end of the current window, so indicate that we
|
||
// want the next window in the sequence next time around. (If we're not up
|
||
// against the end of the window, then we got what we needed and won't be
|
||
// coming around again anyway).
|
||
//
|
||
|
||
SelectNextContigWindow = TRUE;
|
||
WindowRelativeHint = 2;
|
||
}
|
||
else {
|
||
|
||
//
|
||
// FAT 12/16 - we've run up against the end of the volume. Clear the
|
||
// hint, since we now have no idea where to look.
|
||
//
|
||
|
||
WindowRelativeHint = 0;
|
||
}
|
||
#if DBG
|
||
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
|
||
#endif // DBG
|
||
}
|
||
else {
|
||
|
||
if (ExactMatchRequired) {
|
||
|
||
//
|
||
// If our caller required an exact match, then we're hosed. Bail out now.
|
||
//
|
||
|
||
try_leave( Result = FALSE);
|
||
}
|
||
|
||
//
|
||
// Hint failed, drop back to pot luck
|
||
//
|
||
|
||
WindowRelativeHint = 0;
|
||
}
|
||
}
|
||
|
||
if ((0 == WindowRelativeHint) && (0 == ClustersFound)) {
|
||
|
||
if (ClustersRemaining <= Vcb->CurrentWindow->ClustersFree) {
|
||
|
||
//
|
||
// The remaining allocation could be satisfied entirely from this
|
||
// window. We will ask only for what we need, to try and avoid
|
||
// unnecessarily fragmenting large runs of space by always using
|
||
// (part of) the largest run we can find. This call will return the
|
||
// first run large enough.
|
||
//
|
||
|
||
Index = RtlFindClearBits( &Vcb->FreeClusterBitMap, ClustersRemaining, 0);
|
||
|
||
if (-1 != Index) {
|
||
|
||
ClustersFound = ClustersRemaining;
|
||
}
|
||
}
|
||
|
||
if (0 == ClustersFound) {
|
||
|
||
//
|
||
// Still nothing, so just take the largest free run we can find.
|
||
//
|
||
|
||
ClustersFound = RtlFindLongestRunClear( &Vcb->FreeClusterBitMap, &Index );
|
||
|
||
}
|
||
#if DBG
|
||
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
|
||
#endif // DBG
|
||
if (ClustersFound >= ClustersRemaining) {
|
||
|
||
ClustersFound = ClustersRemaining;
|
||
}
|
||
else {
|
||
|
||
//
|
||
// If we just ran up to the end of a window, set up a hint that
|
||
// we'd like the next consecutive window after this one. (FAT32 only)
|
||
//
|
||
|
||
if ( ((Index + ClustersFound) == Vcb->FreeClusterBitMap.SizeOfBitMap) &&
|
||
FatIsFat32( Vcb)
|
||
) {
|
||
|
||
SelectNextContigWindow = TRUE;
|
||
WindowRelativeHint = 2;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
if (ClustersFound == 0) {
|
||
|
||
ULONG FaveWindow = 0;
|
||
BOOLEAN SelectedWindow;
|
||
|
||
//
|
||
// If we found no free clusters on a single-window FAT,
|
||
// there was a bad problem with the free cluster count.
|
||
//
|
||
|
||
if (1 == Vcb->NumberOfWindows) {
|
||
|
||
#ifdef _MSC_VER
|
||
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
|
||
#endif
|
||
FatBugCheck( 0, 5, 0 );
|
||
}
|
||
|
||
//
|
||
// Switch to a new bucket. Possibly the next one if we're
|
||
// currently on a roll (allocating contiguously)
|
||
//
|
||
|
||
SelectedWindow = FALSE;
|
||
|
||
if ( SelectNextContigWindow) {
|
||
|
||
ULONG NextWindow;
|
||
|
||
NextWindow = (((ULONG)((PUCHAR)Vcb->CurrentWindow - (PUCHAR)Vcb->Windows)) / sizeof( FAT_WINDOW)) + 1;
|
||
|
||
if ((NextWindow < Vcb->NumberOfWindows) &&
|
||
( Vcb->Windows[ NextWindow].ClustersFree > 0)
|
||
) {
|
||
|
||
FaveWindow = NextWindow;
|
||
SelectedWindow = TRUE;
|
||
}
|
||
else {
|
||
|
||
if (ExactMatchRequired) {
|
||
|
||
//
|
||
// Some dope tried to allocate a run past the end of the volume...
|
||
//
|
||
|
||
try_leave( Result = FALSE);
|
||
}
|
||
|
||
//
|
||
// Give up on the contiguous allocation attempts
|
||
//
|
||
|
||
WindowRelativeHint = 0;
|
||
}
|
||
|
||
SelectNextContigWindow = FALSE;
|
||
}
|
||
|
||
if (!SelectedWindow) {
|
||
|
||
//
|
||
// Select a new window to begin allocating from
|
||
//
|
||
|
||
FaveWindow = FatSelectBestWindow( Vcb);
|
||
}
|
||
|
||
//
|
||
// By now we'd better have found a window with some free clusters
|
||
//
|
||
|
||
if (0 == Vcb->Windows[ FaveWindow].ClustersFree) {
|
||
|
||
#ifdef _MSC_VER
|
||
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
|
||
#endif
|
||
FatBugCheck( 0, 5, 1 );
|
||
}
|
||
|
||
Wait = BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
SetFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
|
||
FatExamineFatEntries( IrpContext, Vcb,
|
||
0,
|
||
0,
|
||
FALSE,
|
||
&Vcb->Windows[FaveWindow],
|
||
NULL);
|
||
|
||
if (!Wait) {
|
||
|
||
ClearFlag(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT);
|
||
}
|
||
|
||
//
|
||
// Now we'll just go around the loop again, having switched windows,
|
||
// and allocate....
|
||
//
|
||
#if DBG
|
||
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
|
||
#endif //DBG
|
||
} // if (clustersfound == 0)
|
||
else {
|
||
|
||
//
|
||
// Take the clusters we found, convert our index to a cluster number
|
||
// and unlock the bit map.
|
||
//
|
||
|
||
Window = Vcb->CurrentWindow;
|
||
|
||
FatReserveClusters( IrpContext, Vcb, (Index + 2), ClustersFound );
|
||
|
||
Cluster = Index + Window->FirstCluster;
|
||
|
||
Window->ClustersFree -= ClustersFound;
|
||
NT_ASSERT( PreviousClear - ClustersFound == Window->ClustersFree );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
LockedBitMap = FALSE;
|
||
|
||
//
|
||
// Add the newly alloced run to the Mcb.
|
||
//
|
||
|
||
BytesFound = ClustersFound << LogOfBytesPerCluster;
|
||
|
||
FatAddMcbEntry( Vcb, Mcb,
|
||
CurrentVbo,
|
||
FatGetLboFromIndex( Vcb, Cluster ),
|
||
BytesFound );
|
||
|
||
//
|
||
// Connect the last allocated run with this one, and allocate
|
||
// this run on the Fat.
|
||
//
|
||
|
||
if (PriorLastCluster != 0) {
|
||
|
||
FatSetFatEntry( IrpContext,
|
||
Vcb,
|
||
PriorLastCluster,
|
||
(FAT_ENTRY)Cluster );
|
||
}
|
||
|
||
//
|
||
// Update the fat
|
||
//
|
||
|
||
FatAllocateClusters( IrpContext, Vcb, Cluster, ClustersFound );
|
||
|
||
//
|
||
// Prepare for the next iteration.
|
||
//
|
||
|
||
CurrentVbo += BytesFound;
|
||
ClustersRemaining -= ClustersFound;
|
||
PriorLastCluster = Cluster + ClustersFound - 1;
|
||
}
|
||
} // while (clustersremaining)
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatAllocateDiskSpace );
|
||
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
|
||
//
|
||
// Is there any unwinding to do?
|
||
//
|
||
|
||
if ( _SEH2_AbnormalTermination() || (FALSE == Result)) {
|
||
|
||
//
|
||
// Flag to the caller that they're getting nothing
|
||
//
|
||
|
||
*ByteCount = 0;
|
||
|
||
//
|
||
// There are three places we could have taken this exception:
|
||
// when switching the window (FatExamineFatEntries), adding
|
||
// a found run to the Mcb (FatAddMcbEntry), or when writing
|
||
// the changes to the FAT (FatSetFatEntry). In the first case
|
||
// we don't have anything to unwind before deallocation, and
|
||
// can detect this by seeing if we have the ClusterBitmap
|
||
// mutex out.
|
||
|
||
if (!LockedBitMap) {
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
|
||
//
|
||
// In these cases, we have the possiblity that the FAT
|
||
// window is still in place and we need to clear the bits.
|
||
// If the Mcb entry isn't there (we raised trying to add
|
||
// it), the effect of trying to remove it is a noop.
|
||
//
|
||
|
||
if (Window == Vcb->CurrentWindow) {
|
||
|
||
//
|
||
// Cluster reservation works on cluster 2 based window-relative
|
||
// numbers, so we must convert. The subtraction will lose the
|
||
// cluster 2 base, so bias the result.
|
||
//
|
||
|
||
FatUnreserveClusters( IrpContext, Vcb,
|
||
(Cluster - Window->FirstCluster) + 2,
|
||
ClustersFound );
|
||
}
|
||
|
||
//
|
||
// Note that FatDeallocateDiskSpace will take care of adjusting
|
||
// to account for the entries in the Mcb. All we have to account
|
||
// for is the last run that didn't make it.
|
||
//
|
||
|
||
Window->ClustersFree += ClustersFound;
|
||
Vcb->AllocationSupport.NumberOfFreeClusters += ClustersFound;
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
|
||
FatRemoveMcbEntry( Vcb, Mcb, CurrentVbo, BytesFound );
|
||
|
||
} else {
|
||
|
||
//
|
||
// Just drop the mutex now - we didn't manage to do anything
|
||
// that needs to be backed out.
|
||
//
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
}
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Now we have tidied up, we are ready to just send the Mcb
|
||
// off to deallocate disk space
|
||
//
|
||
|
||
FatDeallocateDiskSpace( IrpContext, Vcb, Mcb, FALSE );
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
//
|
||
// Now finally (really), remove all the entries from the mcb
|
||
//
|
||
|
||
FatRemoveMcbEntry( Vcb, Mcb, 0, 0xFFFFFFFF );
|
||
} _SEH2_END;
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatAllocateDiskSpace -> (VOID)\n", 0);
|
||
|
||
} _SEH2_END; // finally
|
||
}
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
|
||
//
|
||
// Limit our zeroing writes to 1 MB.
|
||
//
|
||
|
||
#define MAX_ZERO_MDL_SIZE (1*1024*1024)
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatDeallocateDiskSpace (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN PLARGE_MCB Mcb,
|
||
IN BOOLEAN ZeroOnDeallocate
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This procedure deallocates the disk space denoted by an input
|
||
mcb. Note that the input MCB does not need to necessarily describe
|
||
a chain that ends with a FAT_CLUSTER_LAST entry.
|
||
|
||
Pictorially what is done is the following
|
||
|
||
Fat |--a--|--b--|--c--|
|
||
Mcb |--a--|--b--|--c--|
|
||
|
||
becomes
|
||
|
||
Fat |--0--|--0--|--0--|
|
||
Mcb |--a--|--b--|--c--|
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the VCB being modified
|
||
|
||
Mcb - Supplies the MCB describing the disk space to deallocate. Note
|
||
that Mcb is unchanged by this procedure.
|
||
|
||
|
||
Return Value:
|
||
|
||
None.
|
||
|
||
--*/
|
||
|
||
{
|
||
LBO Lbo;
|
||
VBO Vbo;
|
||
|
||
ULONG RunsInMcb;
|
||
ULONG ByteCount;
|
||
ULONG ClusterCount = 0;
|
||
ULONG ClusterIndex = 0;
|
||
ULONG McbIndex = 0;
|
||
|
||
UCHAR LogOfBytesPerCluster;
|
||
|
||
PFAT_WINDOW Window;
|
||
|
||
NTSTATUS ZeroingStatus = STATUS_SUCCESS;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatDeallocateDiskSpace\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
|
||
|
||
LogOfBytesPerCluster = Vcb->AllocationSupport.LogOfBytesPerCluster;
|
||
|
||
RunsInMcb = FsRtlNumberOfRunsInLargeMcb( Mcb );
|
||
|
||
if ( RunsInMcb == 0 ) {
|
||
|
||
DebugTrace(-1, Dbg, "FatDeallocateDiskSpace -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
//
|
||
// If we are supposed to zero out the allocation before freeing it, do so.
|
||
//
|
||
|
||
if (ZeroOnDeallocate) {
|
||
|
||
_SEH2_TRY {
|
||
|
||
PIRP IoIrp;
|
||
KEVENT IoEvent;
|
||
IO_STATUS_BLOCK Iosb;
|
||
PVOID Buffer = NULL;
|
||
PMDL Mdl;
|
||
ULONG ByteCountToZero;
|
||
ULONG MdlSizeMapped;
|
||
|
||
//
|
||
// Issue the writes down for each run in the Mcb
|
||
//
|
||
|
||
KeInitializeEvent( &IoEvent,
|
||
NotificationEvent,
|
||
FALSE );
|
||
|
||
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
|
||
|
||
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
|
||
|
||
//
|
||
// Assert that Fat files have no holes.
|
||
//
|
||
|
||
NT_ASSERT( Lbo != 0 );
|
||
|
||
//
|
||
// Setup our MDL for the this run.
|
||
//
|
||
|
||
if (ByteCount > MAX_ZERO_MDL_SIZE) {
|
||
Mdl = FatBuildZeroMdl( IrpContext, MAX_ZERO_MDL_SIZE);
|
||
} else {
|
||
Mdl = FatBuildZeroMdl( IrpContext, ByteCount);
|
||
}
|
||
|
||
if (!Mdl) {
|
||
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
|
||
goto try_exit;
|
||
}
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Map the MDL.
|
||
//
|
||
|
||
#ifndef __REACTOS__
|
||
Buffer = MmGetSystemAddressForMdlSafe(Mdl, HighPagePriority|MdlMappingNoExecute);
|
||
#else
|
||
Buffer = MmGetSystemAddressForMdlSafe(Mdl, HighPagePriority);
|
||
#endif
|
||
if (!Buffer) {
|
||
NT_ASSERT( FALSE );
|
||
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
|
||
goto try_exit2;
|
||
}
|
||
|
||
//
|
||
// We might not have not been able to get an MDL big enough to map the whole
|
||
// run. In this case, break up the write.
|
||
//
|
||
|
||
MdlSizeMapped = min( ByteCount, Mdl->ByteCount );
|
||
ByteCountToZero = ByteCount;
|
||
|
||
//
|
||
// Loop until there are no bytes left to write
|
||
//
|
||
|
||
while (ByteCountToZero != 0) {
|
||
|
||
//
|
||
// Write zeros to each run.
|
||
//
|
||
|
||
KeClearEvent( &IoEvent );
|
||
|
||
IoIrp = IoBuildSynchronousFsdRequest( IRP_MJ_WRITE,
|
||
Vcb->TargetDeviceObject,
|
||
Buffer,
|
||
MdlSizeMapped,
|
||
(PLARGE_INTEGER)&Lbo,
|
||
&IoEvent,
|
||
&Iosb );
|
||
|
||
if (IoIrp == NULL) {
|
||
NT_ASSERT( FALSE );
|
||
ZeroingStatus = STATUS_INSUFFICIENT_RESOURCES;
|
||
goto try_exit2;
|
||
}
|
||
|
||
//
|
||
// Set a flag indicating that we want to write through any
|
||
// cache on the controller. This eliminates the need for
|
||
// an explicit flush-device after the write.
|
||
//
|
||
|
||
SetFlag( IoGetNextIrpStackLocation(IoIrp)->Flags, SL_WRITE_THROUGH );
|
||
|
||
ZeroingStatus = IoCallDriver( Vcb->TargetDeviceObject, IoIrp );
|
||
|
||
if (ZeroingStatus == STATUS_PENDING) {
|
||
|
||
(VOID)KeWaitForSingleObject( &IoEvent,
|
||
Executive,
|
||
KernelMode,
|
||
FALSE,
|
||
(PLARGE_INTEGER)NULL );
|
||
|
||
ZeroingStatus = Iosb.Status;
|
||
}
|
||
|
||
if (!NT_SUCCESS( ZeroingStatus )) {
|
||
NT_ASSERT( FALSE );
|
||
goto try_exit2;
|
||
}
|
||
|
||
//
|
||
// Increment the starting offset where we will zero.
|
||
//
|
||
|
||
Lbo += MdlSizeMapped;
|
||
|
||
//
|
||
// Decrement ByteCount
|
||
//
|
||
|
||
ByteCountToZero -= MdlSizeMapped;
|
||
|
||
if (ByteCountToZero < MdlSizeMapped) {
|
||
MdlSizeMapped = ByteCountToZero;
|
||
}
|
||
|
||
}
|
||
|
||
try_exit2:
|
||
|
||
NOTHING;
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
if (!FlagOn( Mdl->MdlFlags, MDL_SOURCE_IS_NONPAGED_POOL) &&
|
||
FlagOn( Mdl->MdlFlags, MDL_MAPPED_TO_SYSTEM_VA )) {
|
||
|
||
MmUnmapLockedPages( Mdl->MappedSystemVa, Mdl );
|
||
}
|
||
IoFreeMdl( Mdl );
|
||
} _SEH2_END;
|
||
|
||
}
|
||
|
||
try_exit:
|
||
|
||
NOTHING;
|
||
|
||
} _SEH2_EXCEPT(FatExceptionFilter( NULL, _SEH2_GetExceptionInformation() )) {
|
||
|
||
//
|
||
// If we failed to zero for some reason, still go ahead and deallocate
|
||
// the clusters. Otherwise we'll leak space from the volume.
|
||
//
|
||
|
||
ZeroingStatus = _SEH2_GetExceptionCode();
|
||
|
||
} _SEH2_END;
|
||
|
||
}
|
||
|
||
NT_ASSERT( NT_SUCCESS(ZeroingStatus) );
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Run though the Mcb, freeing all the runs in the fat.
|
||
//
|
||
// We do this in two steps (first update the fat, then the bitmap
|
||
// (which can't fail)) to prevent other people from taking clusters
|
||
// that we need to re-allocate in the event of unwind.
|
||
//
|
||
|
||
ExAcquireResourceSharedLite(&Vcb->ChangeBitMapResource, TRUE);
|
||
|
||
RunsInMcb = FsRtlNumberOfRunsInLargeMcb( Mcb );
|
||
|
||
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
|
||
|
||
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
|
||
|
||
//
|
||
// Assert that Fat files have no holes.
|
||
//
|
||
|
||
NT_ASSERT( Lbo != 0 );
|
||
|
||
//
|
||
// Write FAT_CLUSTER_AVAILABLE to each cluster in the run.
|
||
//
|
||
|
||
if (ByteCount == 0xFFFFFFFF) {
|
||
|
||
//
|
||
// Special case the computation of ClusterCout
|
||
// when file is of max size (4GiB - 1).
|
||
//
|
||
|
||
ClusterCount = (1 << (32 - LogOfBytesPerCluster));
|
||
|
||
} else {
|
||
|
||
ClusterCount = ByteCount >> LogOfBytesPerCluster;
|
||
}
|
||
|
||
ClusterIndex = FatGetIndexFromLbo( Vcb, Lbo );
|
||
|
||
FatFreeClusters( IrpContext, Vcb, ClusterIndex, ClusterCount );
|
||
}
|
||
|
||
//
|
||
// From now on, nothing can go wrong .... (as in raise)
|
||
//
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
|
||
for ( McbIndex = 0; McbIndex < RunsInMcb; McbIndex++ ) {
|
||
|
||
ULONG ClusterEnd;
|
||
ULONG MyStart, MyLength, count;
|
||
#if DBG
|
||
#ifndef __REACTOS__
|
||
ULONG PreviousClear = 0;
|
||
#endif
|
||
ULONG i = 0;
|
||
#endif
|
||
|
||
FatGetNextMcbEntry( Vcb, Mcb, McbIndex, &Vbo, &Lbo, &ByteCount );
|
||
|
||
//
|
||
// Mark the bits clear in the FreeClusterBitMap.
|
||
//
|
||
|
||
if (ByteCount == 0xFFFFFFFF) {
|
||
|
||
//
|
||
// Special case the computation of ClusterCout
|
||
// when file is of max size (2^32 - 1).
|
||
//
|
||
|
||
ClusterCount = (1 << (32 - LogOfBytesPerCluster));
|
||
|
||
} else {
|
||
|
||
ClusterCount = ByteCount >> LogOfBytesPerCluster;
|
||
}
|
||
|
||
ClusterIndex = FatGetIndexFromLbo( Vcb, Lbo );
|
||
|
||
Window = Vcb->CurrentWindow;
|
||
|
||
//
|
||
// If we've divided the bitmap, elide bitmap manipulation for
|
||
// runs that are outside the current bucket.
|
||
//
|
||
|
||
ClusterEnd = ClusterIndex + ClusterCount - 1;
|
||
|
||
if (!(ClusterIndex > Window->LastCluster ||
|
||
ClusterEnd < Window->FirstCluster)) {
|
||
|
||
//
|
||
// The run being freed overlaps the current bucket, so we'll
|
||
// have to clear some bits.
|
||
//
|
||
|
||
if (ClusterIndex < Window->FirstCluster &&
|
||
ClusterEnd > Window->LastCluster) {
|
||
|
||
MyStart = Window->FirstCluster;
|
||
MyLength = Window->LastCluster - Window->FirstCluster + 1;
|
||
|
||
} else if (ClusterIndex < Window->FirstCluster) {
|
||
|
||
MyStart = Window->FirstCluster;
|
||
MyLength = ClusterEnd - Window->FirstCluster + 1;
|
||
|
||
} else {
|
||
|
||
//
|
||
// The range being freed starts in the bucket, and may possibly
|
||
// extend beyond the bucket.
|
||
//
|
||
|
||
MyStart = ClusterIndex;
|
||
|
||
if (ClusterEnd <= Window->LastCluster) {
|
||
|
||
MyLength = ClusterCount;
|
||
|
||
} else {
|
||
|
||
MyLength = Window->LastCluster - ClusterIndex + 1;
|
||
}
|
||
}
|
||
|
||
if (MyLength == 0) {
|
||
|
||
continue;
|
||
}
|
||
|
||
#if DBG
|
||
#ifndef __REACTOS__
|
||
#ifdef _MSC_VER
|
||
#pragma prefast( suppress:28931, "this is DBG build only" )
|
||
#endif
|
||
PreviousClear = RtlNumberOfClearBits( &Vcb->FreeClusterBitMap );
|
||
#endif
|
||
|
||
|
||
//
|
||
// Verify that the Bits are all really set.
|
||
//
|
||
|
||
NT_ASSERT( MyStart + MyLength - Window->FirstCluster <= Vcb->FreeClusterBitMap.SizeOfBitMap );
|
||
|
||
for (i = 0; i < MyLength; i++) {
|
||
|
||
NT_ASSERT( RtlCheckBit(&Vcb->FreeClusterBitMap,
|
||
MyStart - Window->FirstCluster + i) == 1 );
|
||
}
|
||
#endif // DBG
|
||
|
||
FatUnreserveClusters( IrpContext, Vcb,
|
||
MyStart - Window->FirstCluster + 2,
|
||
MyLength );
|
||
}
|
||
|
||
//
|
||
// Adjust the ClustersFree count for each bitmap window, even the ones
|
||
// that are not the current window.
|
||
//
|
||
|
||
if (FatIsFat32(Vcb)) {
|
||
|
||
Window = &Vcb->Windows[FatWindowOfCluster( ClusterIndex )];
|
||
|
||
} else {
|
||
|
||
Window = &Vcb->Windows[0];
|
||
}
|
||
|
||
MyStart = ClusterIndex;
|
||
|
||
for (MyLength = ClusterCount; MyLength > 0; MyLength -= count) {
|
||
|
||
count = FatMin(Window->LastCluster - MyStart + 1, MyLength);
|
||
Window->ClustersFree += count;
|
||
|
||
//
|
||
// If this was not the last window this allocation spanned,
|
||
// advance to the next.
|
||
//
|
||
|
||
if (MyLength != count) {
|
||
|
||
Window++;
|
||
MyStart = Window->FirstCluster;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Deallocation is now complete. Adjust the free cluster count.
|
||
//
|
||
|
||
Vcb->AllocationSupport.NumberOfFreeClusters += ClusterCount;
|
||
}
|
||
|
||
#if DBG
|
||
if (Vcb->CurrentWindow->ClustersFree !=
|
||
RtlNumberOfClearBits(&Vcb->FreeClusterBitMap)) {
|
||
|
||
DbgPrint("%x vs %x\n", Vcb->CurrentWindow->ClustersFree,
|
||
RtlNumberOfClearBits(&Vcb->FreeClusterBitMap));
|
||
|
||
DbgPrint("%x for %x\n", ClusterIndex, ClusterCount);
|
||
}
|
||
#endif
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatDeallocateDiskSpace );
|
||
|
||
//
|
||
// Is there any unwinding to do?
|
||
//
|
||
|
||
ExReleaseResourceLite(&Vcb->ChangeBitMapResource);
|
||
|
||
if ( _SEH2_AbnormalTermination() ) {
|
||
|
||
LBO LocalLbo;
|
||
VBO LocalVbo;
|
||
|
||
ULONG Index;
|
||
ULONG Clusters;
|
||
ULONG FatIndex;
|
||
ULONG PriorLastIndex;
|
||
|
||
//
|
||
// For each entry we already deallocated, reallocate it,
|
||
// chaining together as nessecary. Note that we continue
|
||
// up to and including the last "for" iteration even though
|
||
// the SetFatRun could not have been successful. This
|
||
// allows us a convienent way to re-link the final successful
|
||
// SetFatRun.
|
||
//
|
||
// It is possible that the reason we got here will prevent us
|
||
// from succeeding in this operation.
|
||
//
|
||
|
||
PriorLastIndex = 0;
|
||
|
||
for (Index = 0; Index <= McbIndex; Index++) {
|
||
|
||
FatGetNextMcbEntry(Vcb, Mcb, Index, &LocalVbo, &LocalLbo, &ByteCount);
|
||
|
||
if (ByteCount == 0xFFFFFFFF) {
|
||
|
||
//
|
||
// Special case the computation of ClusterCout
|
||
// when file is of max size (2^32 - 1).
|
||
//
|
||
|
||
Clusters = (1 << (32 - LogOfBytesPerCluster));
|
||
|
||
} else {
|
||
|
||
Clusters = ByteCount >> LogOfBytesPerCluster;
|
||
}
|
||
|
||
FatIndex = FatGetIndexFromLbo( Vcb, LocalLbo );
|
||
|
||
//
|
||
// We must always restore the prior iteration's last
|
||
// entry, pointing it to the first cluster of this run.
|
||
//
|
||
|
||
if (PriorLastIndex != 0) {
|
||
|
||
FatSetFatEntry( IrpContext,
|
||
Vcb,
|
||
PriorLastIndex,
|
||
(FAT_ENTRY)FatIndex );
|
||
}
|
||
|
||
//
|
||
// If this is not the last entry (the one that failed)
|
||
// then reallocate the disk space on the fat.
|
||
//
|
||
|
||
if ( Index < McbIndex ) {
|
||
|
||
FatAllocateClusters(IrpContext, Vcb, FatIndex, Clusters);
|
||
|
||
PriorLastIndex = FatIndex + Clusters - 1;
|
||
}
|
||
}
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatDeallocateDiskSpace -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatSplitAllocation (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN OUT PLARGE_MCB Mcb,
|
||
IN VBO SplitAtVbo,
|
||
OUT PLARGE_MCB RemainingMcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This procedure takes a single mcb and splits its allocation into
|
||
two separate allocation units. The separation must only be done
|
||
on cluster boundaries, otherwise we bugcheck.
|
||
|
||
On the disk this actually works by inserting a FAT_CLUSTER_LAST into
|
||
the last index of the first part being split out.
|
||
|
||
Pictorially what is done is the following (where ! denotes the end of
|
||
the fat chain (i.e., FAT_CLUSTER_LAST)):
|
||
|
||
|
||
Mcb |--a--|--b--|--c--|--d--|--e--|--f--|
|
||
|
||
^
|
||
SplitAtVbo ---------------------+
|
||
|
||
RemainingMcb (empty)
|
||
|
||
becomes
|
||
|
||
Mcb |--a--|--b--|--c--!
|
||
|
||
|
||
RemainingMcb |--d--|--e--|--f--|
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the VCB being modified
|
||
|
||
Mcb - Supplies the MCB describing the allocation being split into
|
||
two parts. Upon return this Mcb now contains the first chain.
|
||
|
||
SplitAtVbo - Supplies the VBO of the first byte for the second chain
|
||
that we creating.
|
||
|
||
RemainingMcb - Receives the MCB describing the second chain of allocated
|
||
disk space. The caller passes in an initialized Mcb that
|
||
is filled in by this procedure STARTING AT VBO 0.
|
||
|
||
Return Value:
|
||
|
||
VOID - TRUE if the operation completed and FALSE if it had to
|
||
block but could not.
|
||
|
||
--*/
|
||
|
||
{
|
||
VBO SourceVbo;
|
||
VBO TargetVbo;
|
||
VBO DontCare;
|
||
|
||
LBO Lbo;
|
||
|
||
ULONG ByteCount;
|
||
|
||
#if DBG
|
||
ULONG BytesPerCluster;
|
||
#endif
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatSplitAllocation\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
|
||
DebugTrace( 0, Dbg, " SplitAtVbo = %8lx\n", SplitAtVbo);
|
||
DebugTrace( 0, Dbg, " RemainingMcb = %p\n", RemainingMcb);
|
||
|
||
#if DBG
|
||
BytesPerCluster = 1 << Vcb->AllocationSupport.LogOfBytesPerCluster;
|
||
#endif
|
||
|
||
//
|
||
// Assert that the split point is cluster alligned
|
||
//
|
||
|
||
NT_ASSERT( (SplitAtVbo & (BytesPerCluster - 1)) == 0 );
|
||
|
||
//
|
||
// We should never be handed an empty source MCB and asked to split
|
||
// at a non zero point.
|
||
//
|
||
|
||
NT_ASSERT( !((0 != SplitAtVbo) && (0 == FsRtlNumberOfRunsInLargeMcb( Mcb))));
|
||
|
||
//
|
||
// Assert we were given an empty target Mcb.
|
||
//
|
||
|
||
//
|
||
// This assert is commented out to avoid hitting in the Ea error
|
||
// path. In that case we will be using the same Mcb's to split the
|
||
// allocation that we used to merge them. The target Mcb will contain
|
||
// the runs that the split will attempt to insert.
|
||
//
|
||
//
|
||
// NT_ASSERT( FsRtlNumberOfRunsInMcb( RemainingMcb ) == 0 );
|
||
//
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Move the runs after SplitAtVbo from the souce to the target
|
||
//
|
||
|
||
SourceVbo = SplitAtVbo;
|
||
TargetVbo = 0;
|
||
|
||
while (FatLookupMcbEntry(Vcb, Mcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
|
||
|
||
FatAddMcbEntry( Vcb, RemainingMcb, TargetVbo, Lbo, ByteCount );
|
||
|
||
FatRemoveMcbEntry( Vcb, Mcb, SourceVbo, ByteCount );
|
||
|
||
TargetVbo += ByteCount;
|
||
SourceVbo += ByteCount;
|
||
|
||
//
|
||
// If SourceVbo overflows, we were actually snipping off the end
|
||
// of the maximal file ... and are now done.
|
||
//
|
||
|
||
if (SourceVbo == 0) {
|
||
|
||
break;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Mark the last pre-split cluster as a FAT_LAST_CLUSTER
|
||
//
|
||
|
||
if ( SplitAtVbo != 0 ) {
|
||
|
||
FatLookupLastMcbEntry( Vcb, Mcb, &DontCare, &Lbo, NULL );
|
||
|
||
FatSetFatEntry( IrpContext,
|
||
Vcb,
|
||
FatGetIndexFromLbo( Vcb, Lbo ),
|
||
FAT_CLUSTER_LAST );
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatSplitAllocation );
|
||
|
||
//
|
||
// If we got an exception, we must glue back together the Mcbs
|
||
//
|
||
|
||
if ( _SEH2_AbnormalTermination() ) {
|
||
|
||
TargetVbo = SplitAtVbo;
|
||
SourceVbo = 0;
|
||
|
||
while (FatLookupMcbEntry(Vcb, RemainingMcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
|
||
|
||
FatAddMcbEntry( Vcb, Mcb, TargetVbo, Lbo, ByteCount );
|
||
|
||
FatRemoveMcbEntry( Vcb, RemainingMcb, SourceVbo, ByteCount );
|
||
|
||
TargetVbo += ByteCount;
|
||
SourceVbo += ByteCount;
|
||
}
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatSplitAllocation -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatMergeAllocation (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN OUT PLARGE_MCB Mcb,
|
||
IN PLARGE_MCB SecondMcb
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine takes two separate allocations described by two MCBs and
|
||
joins them together into one allocation.
|
||
|
||
Pictorially what is done is the following (where ! denotes the end of
|
||
the fat chain (i.e., FAT_CLUSTER_LAST)):
|
||
|
||
|
||
Mcb |--a--|--b--|--c--!
|
||
|
||
SecondMcb |--d--|--e--|--f--|
|
||
|
||
becomes
|
||
|
||
Mcb |--a--|--b--|--c--|--d--|--e--|--f--|
|
||
|
||
SecondMcb |--d--|--e--|--f--|
|
||
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the VCB being modified
|
||
|
||
Mcb - Supplies the MCB of the first allocation that is being modified.
|
||
Upon return this Mcb will also describe the newly enlarged
|
||
allocation
|
||
|
||
SecondMcb - Supplies the ZERO VBO BASED MCB of the second allocation
|
||
that is being appended to the first allocation. This
|
||
procedure leaves SecondMcb unchanged.
|
||
|
||
Return Value:
|
||
|
||
VOID - TRUE if the operation completed and FALSE if it had to
|
||
block but could not.
|
||
|
||
--*/
|
||
|
||
{
|
||
VBO SpliceVbo = 0;
|
||
LBO SpliceLbo;
|
||
|
||
VBO SourceVbo;
|
||
VBO TargetVbo = 0;
|
||
|
||
LBO Lbo;
|
||
|
||
ULONG ByteCount;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatMergeAllocation\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " Mcb = %p\n", Mcb);
|
||
DebugTrace( 0, Dbg, " SecondMcb = %p\n", SecondMcb);
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Append the runs from SecondMcb to Mcb
|
||
//
|
||
|
||
(void)FatLookupLastMcbEntry( Vcb, Mcb, &SpliceVbo, &SpliceLbo, NULL );
|
||
|
||
SourceVbo = 0;
|
||
TargetVbo = SpliceVbo + 1;
|
||
|
||
while (FatLookupMcbEntry(Vcb, SecondMcb, SourceVbo, &Lbo, &ByteCount, NULL)) {
|
||
|
||
FatAddMcbEntry( Vcb, Mcb, TargetVbo, Lbo, ByteCount );
|
||
|
||
SourceVbo += ByteCount;
|
||
TargetVbo += ByteCount;
|
||
}
|
||
|
||
//
|
||
// Link the last pre-merge cluster to the first cluster of SecondMcb
|
||
//
|
||
|
||
FatLookupMcbEntry( Vcb, SecondMcb, 0, &Lbo, (PULONG)NULL, NULL );
|
||
|
||
FatSetFatEntry( IrpContext,
|
||
Vcb,
|
||
FatGetIndexFromLbo( Vcb, SpliceLbo ),
|
||
(FAT_ENTRY)FatGetIndexFromLbo( Vcb, Lbo ) );
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatMergeAllocation );
|
||
|
||
//
|
||
// If we got an exception, we must remove the runs added to Mcb
|
||
//
|
||
|
||
if ( _SEH2_AbnormalTermination() ) {
|
||
|
||
ULONG CutLength;
|
||
|
||
if ((CutLength = TargetVbo - (SpliceVbo + 1)) != 0) {
|
||
|
||
FatRemoveMcbEntry( Vcb, Mcb, SpliceVbo + 1, CutLength);
|
||
}
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatMergeAllocation -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
//
|
||
// Internal support routine
|
||
//
|
||
|
||
CLUSTER_TYPE
|
||
FatInterpretClusterType (
|
||
IN PVCB Vcb,
|
||
IN FAT_ENTRY Entry
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This procedure tells the caller how to interpret the input fat table
|
||
entry. It will indicate if the fat cluster is available, resereved,
|
||
bad, the last one, or the another fat index. This procedure can deal
|
||
with both 12 and 16 bit fat.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to examine, yields 12/16 bit info
|
||
|
||
Entry - Supplies the fat entry to examine
|
||
|
||
Return Value:
|
||
|
||
CLUSTER_TYPE - Is the type of the input Fat entry
|
||
|
||
--*/
|
||
|
||
{
|
||
DebugTrace(+1, Dbg, "InterpretClusterType\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " Entry = %8lx\n", Entry);
|
||
|
||
PAGED_CODE();
|
||
|
||
switch(Vcb->AllocationSupport.FatIndexBitSize ) {
|
||
case 32:
|
||
Entry &= FAT32_ENTRY_MASK;
|
||
break;
|
||
|
||
case 12:
|
||
NT_ASSERT( Entry <= 0xfff );
|
||
if (Entry >= 0x0ff0) {
|
||
Entry |= 0x0FFFF000;
|
||
}
|
||
break;
|
||
|
||
default:
|
||
case 16:
|
||
NT_ASSERT( Entry <= 0xffff );
|
||
if (Entry >= 0x0fff0) {
|
||
Entry |= 0x0FFF0000;
|
||
}
|
||
break;
|
||
}
|
||
|
||
if (Entry == FAT_CLUSTER_AVAILABLE) {
|
||
|
||
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterAvailable\n", 0);
|
||
|
||
return FatClusterAvailable;
|
||
|
||
} else if (Entry < FAT_CLUSTER_RESERVED) {
|
||
|
||
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterNext\n", 0);
|
||
|
||
return FatClusterNext;
|
||
|
||
} else if (Entry < FAT_CLUSTER_BAD) {
|
||
|
||
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterReserved\n", 0);
|
||
|
||
return FatClusterReserved;
|
||
|
||
} else if (Entry == FAT_CLUSTER_BAD) {
|
||
|
||
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterBad\n", 0);
|
||
|
||
return FatClusterBad;
|
||
|
||
} else {
|
||
|
||
DebugTrace(-1, Dbg, "FatInterpretClusterType -> FatClusterLast\n", 0);
|
||
|
||
return FatClusterLast;
|
||
}
|
||
}
|
||
|
||
|
||
//
|
||
// Internal support routine
|
||
//
|
||
|
||
VOID
|
||
FatLookupFatEntry (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG FatIndex,
|
||
IN OUT PULONG FatEntry,
|
||
IN OUT PFAT_ENUMERATION_CONTEXT Context
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine takes an index into the fat and gives back the value
|
||
in the Fat at this index. At any given time, for a 16 bit fat, this
|
||
routine allows only one page per volume of the fat to be pinned in
|
||
memory. For a 12 bit bit fat, the entire fat (max 6k) is pinned. This
|
||
extra layer of caching makes the vast majority of requests very
|
||
fast. The context for this caching stored in a structure in the Vcb.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to examine, yields 12/16 bit info,
|
||
fat access context, etc.
|
||
|
||
FatIndex - Supplies the fat index to examine.
|
||
|
||
FatEntry - Receives the fat entry pointed to by FatIndex. Note that
|
||
it must point to non-paged pool.
|
||
|
||
Context - This structure keeps track of a page of pinned fat between calls.
|
||
|
||
--*/
|
||
|
||
{
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatLookupFatEntry\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " FatIndex = %4x\n", FatIndex);
|
||
DebugTrace( 0, Dbg, " FatEntry = %8lx\n", FatEntry);
|
||
|
||
//
|
||
// Make sure they gave us a valid fat index.
|
||
//
|
||
|
||
FatVerifyIndexIsValid(IrpContext, Vcb, FatIndex);
|
||
|
||
//
|
||
// Case on 12 or 16 bit fats.
|
||
//
|
||
// In the 12 bit case (mostly floppies) we always have the whole fat
|
||
// (max 6k bytes) pinned during allocation operations. This is possibly
|
||
// a wee bit slower, but saves headaches over fat entries with 8 bits
|
||
// on one page, and 4 bits on the next.
|
||
//
|
||
// The 16 bit case always keeps the last used page pinned until all
|
||
// operations are done and it is unpinned.
|
||
//
|
||
|
||
//
|
||
// DEAL WITH 12 BIT CASE
|
||
//
|
||
|
||
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
|
||
|
||
//
|
||
// Check to see if the fat is already pinned, otherwise pin it.
|
||
//
|
||
|
||
if (Context->Bcb == NULL) {
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
FatReservedBytes( &Vcb->Bpb ),
|
||
FatBytesPerFat( &Vcb->Bpb ),
|
||
&Context->Bcb,
|
||
&Context->PinnedPage );
|
||
}
|
||
|
||
//
|
||
// Load the return value.
|
||
//
|
||
|
||
|
||
FatLookup12BitEntry( Context->PinnedPage, FatIndex, FatEntry );
|
||
|
||
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
|
||
|
||
//
|
||
// DEAL WITH 32 BIT CASE
|
||
//
|
||
|
||
ULONG PageEntryOffset;
|
||
ULONG OffsetIntoVolumeFile;
|
||
|
||
//
|
||
// Initialize two local variables that help us.
|
||
//
|
||
OffsetIntoVolumeFile = FatReservedBytes(&Vcb->Bpb) + FatIndex * sizeof(FAT_ENTRY);
|
||
PageEntryOffset = (OffsetIntoVolumeFile % PAGE_SIZE) / sizeof(FAT_ENTRY);
|
||
|
||
//
|
||
// Check to see if we need to read in a new page of fat
|
||
//
|
||
|
||
if ((Context->Bcb == NULL) ||
|
||
(OffsetIntoVolumeFile / PAGE_SIZE != Context->VboOfPinnedPage / PAGE_SIZE)) {
|
||
|
||
//
|
||
// The entry wasn't in the pinned page, so must we unpin the current
|
||
// page (if any) and read in a new page.
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Context->Bcb );
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
OffsetIntoVolumeFile & ~(PAGE_SIZE - 1),
|
||
PAGE_SIZE,
|
||
&Context->Bcb,
|
||
&Context->PinnedPage );
|
||
|
||
Context->VboOfPinnedPage = OffsetIntoVolumeFile & ~(PAGE_SIZE - 1);
|
||
}
|
||
|
||
//
|
||
// Grab the fat entry from the pinned page, and return
|
||
//
|
||
|
||
*FatEntry = ((PULONG)(Context->PinnedPage))[PageEntryOffset] & FAT32_ENTRY_MASK;
|
||
|
||
} else {
|
||
|
||
//
|
||
// DEAL WITH 16 BIT CASE
|
||
//
|
||
|
||
ULONG PageEntryOffset;
|
||
ULONG OffsetIntoVolumeFile;
|
||
|
||
//
|
||
// Initialize two local variables that help us.
|
||
//
|
||
|
||
OffsetIntoVolumeFile = FatReservedBytes(&Vcb->Bpb) + FatIndex * sizeof(USHORT);
|
||
PageEntryOffset = (OffsetIntoVolumeFile % PAGE_SIZE) / sizeof(USHORT);
|
||
|
||
//
|
||
// Check to see if we need to read in a new page of fat
|
||
//
|
||
|
||
if ((Context->Bcb == NULL) ||
|
||
(OffsetIntoVolumeFile / PAGE_SIZE != Context->VboOfPinnedPage / PAGE_SIZE)) {
|
||
|
||
//
|
||
// The entry wasn't in the pinned page, so must we unpin the current
|
||
// page (if any) and read in a new page.
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Context->Bcb );
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
OffsetIntoVolumeFile & ~(PAGE_SIZE - 1),
|
||
PAGE_SIZE,
|
||
&Context->Bcb,
|
||
&Context->PinnedPage );
|
||
|
||
Context->VboOfPinnedPage = OffsetIntoVolumeFile & ~(PAGE_SIZE - 1);
|
||
}
|
||
|
||
//
|
||
// Grab the fat entry from the pinned page, and return
|
||
//
|
||
|
||
*FatEntry = ((PUSHORT)(Context->PinnedPage))[PageEntryOffset];
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatLookupFatEntry -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
|
||
_Requires_lock_held_(_Global_critical_region_)
|
||
VOID
|
||
FatSetFatEntry (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG FatIndex,
|
||
IN FAT_ENTRY FatEntry
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine takes an index into the fat and puts a value in the Fat
|
||
at this index. The routine special cases 12, 16 and 32 bit fats. In
|
||
all cases we go to the cache manager for a piece of the fat.
|
||
|
||
We have a special form of this call for setting the DOS-style dirty bit.
|
||
Unlike the dirty bit in the boot sector, we do not go to special effort
|
||
to make sure that this hits the disk synchronously - if the system goes
|
||
down in the window between the dirty bit being set in the boot sector
|
||
and the FAT index zero dirty bit being lazy written, then life is tough.
|
||
|
||
The only possible scenario is that Win9x may see what it thinks is a clean
|
||
volume that really isn't (hopefully Memphis will pay attention to our dirty
|
||
bit as well). The dirty bit will get out quickly, and if heavy activity is
|
||
occurring, then the dirty bit should actually be there virtually all of the
|
||
time since the act of cleaning the volume is the "rare" occurance.
|
||
|
||
There are synchronization concerns that would crop up if we tried to make
|
||
this synchronous. This thread may already own the Bcb shared for the first
|
||
sector of the FAT (so we can't get it exclusive for a writethrough). This
|
||
would require some more serious replumbing to work around than I want to
|
||
consider at this time.
|
||
|
||
We can and do, however, synchronously set the bit clean.
|
||
|
||
At this point the reader should understand why the NT dirty bit is where it is.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to examine, yields 12/16/32 bit info, etc.
|
||
|
||
FatIndex - Supplies the destination fat index.
|
||
|
||
FatEntry - Supplies the source fat entry.
|
||
|
||
--*/
|
||
|
||
{
|
||
LBO Lbo;
|
||
PBCB Bcb = NULL;
|
||
ULONG SectorSize;
|
||
ULONG OffsetIntoVolumeFile;
|
||
ULONG WasWait = TRUE;
|
||
BOOLEAN RegularOperation = TRUE;
|
||
BOOLEAN CleaningOperation = FALSE;
|
||
BOOLEAN ReleaseMutex = FALSE;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatSetFatEntry\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " FatIndex = %4x\n", FatIndex);
|
||
DebugTrace( 0, Dbg, " FatEntry = %4x\n", FatEntry);
|
||
|
||
//
|
||
// Make sure they gave us a valid fat index if this isn't the special
|
||
// clean-bit modifying call.
|
||
//
|
||
|
||
if (FatIndex == FAT_DIRTY_BIT_INDEX) {
|
||
|
||
//
|
||
// We are setting the clean bit state. Of course, we could
|
||
// have corruption that would cause us to try to fiddle the
|
||
// reserved index - we guard against this by having the
|
||
// special entry values use the reserved high 4 bits that
|
||
// we know that we'll never try to set.
|
||
//
|
||
|
||
//
|
||
// We don't want to repin the FAT pages involved here. Just
|
||
// let the lazy writer hit them when it can.
|
||
//
|
||
|
||
RegularOperation = FALSE;
|
||
|
||
switch (FatEntry) {
|
||
case FAT_CLEAN_VOLUME:
|
||
FatEntry = (FAT_ENTRY)FAT_CLEAN_ENTRY;
|
||
CleaningOperation = TRUE;
|
||
break;
|
||
|
||
case FAT_DIRTY_VOLUME:
|
||
switch (Vcb->AllocationSupport.FatIndexBitSize) {
|
||
case 12:
|
||
FatEntry = FAT12_DIRTY_ENTRY;
|
||
break;
|
||
|
||
case 32:
|
||
FatEntry = FAT32_DIRTY_ENTRY;
|
||
break;
|
||
|
||
default:
|
||
FatEntry = FAT16_DIRTY_ENTRY;
|
||
break;
|
||
}
|
||
break;
|
||
|
||
default:
|
||
FatRaiseStatus( IrpContext, STATUS_FILE_CORRUPT_ERROR );
|
||
break;
|
||
}
|
||
|
||
//
|
||
// Disable dirtying semantics for the duration of this operation. Force this
|
||
// operation to wait for the duration.
|
||
//
|
||
|
||
WasWait = FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT );
|
||
SetFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT | IRP_CONTEXT_FLAG_DISABLE_DIRTY );
|
||
|
||
} else {
|
||
|
||
NT_ASSERT( !(FatEntry & ~FAT32_ENTRY_MASK) );
|
||
FatVerifyIndexIsValid(IrpContext, Vcb, FatIndex);
|
||
}
|
||
|
||
//
|
||
// Set Sector Size
|
||
//
|
||
|
||
SectorSize = 1 << Vcb->AllocationSupport.LogOfBytesPerSector;
|
||
|
||
//
|
||
// Case on 12 or 16 bit fats.
|
||
//
|
||
// In the 12 bit case (mostly floppies) we always have the whole fat
|
||
// (max 6k bytes) pinned during allocation operations. This is possibly
|
||
// a wee bit slower, but saves headaches over fat entries with 8 bits
|
||
// on one page, and 4 bits on the next.
|
||
//
|
||
// In the 16 bit case we only read the page that we need to set the fat
|
||
// entry.
|
||
//
|
||
|
||
//
|
||
// DEAL WITH 12 BIT CASE
|
||
//
|
||
|
||
_SEH2_TRY {
|
||
|
||
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
|
||
|
||
PVOID PinnedFat;
|
||
|
||
//
|
||
// Make sure we have a valid entry
|
||
//
|
||
|
||
FatEntry &= 0xfff;
|
||
|
||
//
|
||
// We read in the entire fat. Note that using prepare write marks
|
||
// the bcb pre-dirty, so we don't have to do it explicitly.
|
||
//
|
||
|
||
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) + FatIndex * 3 / 2;
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
FatReservedBytes( &Vcb->Bpb ),
|
||
FatBytesPerFat( &Vcb->Bpb ),
|
||
&Bcb,
|
||
&PinnedFat,
|
||
RegularOperation,
|
||
FALSE );
|
||
|
||
//
|
||
// Mark the sector(s) dirty in the DirtyFatMcb. This call is
|
||
// complicated somewhat for the 12 bit case since a single
|
||
// entry write can span two sectors (and pages).
|
||
//
|
||
// Get the Lbo for the sector where the entry starts, and add it to
|
||
// the dirty fat Mcb.
|
||
//
|
||
|
||
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
|
||
|
||
//
|
||
// If the entry started on the last byte of the sector, it continues
|
||
// to the next sector, so mark the next sector dirty as well.
|
||
//
|
||
// Note that this entry will simply coalese with the last entry,
|
||
// so this operation cannot fail. Also if we get this far, we have
|
||
// made it, so no unwinding will be needed.
|
||
//
|
||
|
||
if ( (OffsetIntoVolumeFile & (SectorSize - 1)) == (SectorSize - 1) ) {
|
||
|
||
Lbo += SectorSize;
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
|
||
}
|
||
|
||
//
|
||
// Store the entry into the fat; we need a little synchonization
|
||
// here and can't use a spinlock since the bytes might not be
|
||
// resident.
|
||
//
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
|
||
FatSet12BitEntry( PinnedFat, FatIndex, FatEntry );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
|
||
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
|
||
|
||
//
|
||
// DEAL WITH 32 BIT CASE
|
||
//
|
||
|
||
PULONG PinnedFatEntry32;
|
||
|
||
//
|
||
// Read in a new page of fat
|
||
//
|
||
|
||
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) +
|
||
FatIndex * sizeof( FAT_ENTRY );
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
OffsetIntoVolumeFile,
|
||
sizeof(FAT_ENTRY),
|
||
&Bcb,
|
||
(PVOID *)&PinnedFatEntry32,
|
||
RegularOperation,
|
||
FALSE );
|
||
//
|
||
// Mark the sector dirty in the DirtyFatMcb
|
||
//
|
||
|
||
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
|
||
|
||
//
|
||
// Store the FatEntry to the pinned page.
|
||
//
|
||
// Preserve the reserved bits in FAT32 entries in the file heap.
|
||
//
|
||
|
||
#ifdef ALPHA
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
#endif // ALPHA
|
||
|
||
if (FatIndex != FAT_DIRTY_BIT_INDEX) {
|
||
|
||
*PinnedFatEntry32 = ((*PinnedFatEntry32 & ~FAT32_ENTRY_MASK) | FatEntry);
|
||
|
||
} else {
|
||
|
||
*PinnedFatEntry32 = FatEntry;
|
||
}
|
||
|
||
#ifdef ALPHA
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
#endif // ALPHA
|
||
|
||
} else {
|
||
|
||
//
|
||
// DEAL WITH 16 BIT CASE
|
||
//
|
||
|
||
PUSHORT PinnedFatEntry;
|
||
|
||
//
|
||
// Read in a new page of fat
|
||
//
|
||
|
||
OffsetIntoVolumeFile = FatReservedBytes( &Vcb->Bpb ) +
|
||
FatIndex * sizeof(USHORT);
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
OffsetIntoVolumeFile,
|
||
sizeof(USHORT),
|
||
&Bcb,
|
||
(PVOID *)&PinnedFatEntry,
|
||
RegularOperation,
|
||
FALSE );
|
||
//
|
||
// Mark the sector dirty in the DirtyFatMcb
|
||
//
|
||
|
||
Lbo = OffsetIntoVolumeFile & ~(SectorSize - 1);
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize);
|
||
|
||
//
|
||
// Store the FatEntry to the pinned page.
|
||
//
|
||
// We need extra synchronization here for broken architectures
|
||
// like the ALPHA that don't support atomic 16 bit writes.
|
||
//
|
||
|
||
#ifdef ALPHA
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
#endif // ALPHA
|
||
|
||
*PinnedFatEntry = (USHORT)FatEntry;
|
||
|
||
#ifdef ALPHA
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
#endif // ALPHA
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
DebugUnwind( FatSetFatEntry );
|
||
|
||
//
|
||
// Re-enable volume dirtying in case this was a dirty bit operation.
|
||
//
|
||
|
||
ClearFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_DIRTY );
|
||
|
||
//
|
||
// Make this operation asynchronous again if needed.
|
||
//
|
||
|
||
if (!WasWait) {
|
||
|
||
ClearFlag( IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT );
|
||
}
|
||
|
||
//
|
||
// If we still somehow have the Mutex, release it.
|
||
//
|
||
|
||
if (ReleaseMutex) {
|
||
|
||
NT_ASSERT( _SEH2_AbnormalTermination() );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
}
|
||
|
||
//
|
||
// Unpin the Bcb. For cleaning operations or if the corruption was detected while mounting we make this write-through.
|
||
//
|
||
|
||
if ((CleaningOperation ||
|
||
FlagOn(Vcb->VcbState, VCB_STATE_FLAG_MOUNT_IN_PROGRESS)) &&
|
||
Bcb) {
|
||
|
||
IO_STATUS_BLOCK IgnoreStatus;
|
||
|
||
CcRepinBcb( Bcb );
|
||
CcUnpinData( Bcb );
|
||
DbgDoit( IrpContext->PinCount -= 1 );
|
||
CcUnpinRepinnedBcb( Bcb, TRUE, &IgnoreStatus );
|
||
|
||
} else {
|
||
|
||
FatUnpinBcb(IrpContext, Bcb);
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatSetFatEntry -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
//
|
||
// Internal support routine
|
||
//
|
||
|
||
VOID
|
||
FatSetFatRun (
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG StartingFatIndex,
|
||
IN ULONG ClusterCount,
|
||
IN BOOLEAN ChainTogether
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine sets a continuous run of clusters in the fat. If ChainTogether
|
||
is TRUE, then the clusters are linked together as in normal Fat fasion,
|
||
with the last cluster receiving FAT_CLUSTER_LAST. If ChainTogether is
|
||
FALSE, all the entries are set to FAT_CLUSTER_AVAILABLE, effectively
|
||
freeing all the clusters in the run.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the Vcb to examine, yields 12/16 bit info, etc.
|
||
|
||
StartingFatIndex - Supplies the destination fat index.
|
||
|
||
ClusterCount - Supplies the number of contiguous clusters to work on.
|
||
|
||
ChainTogether - Tells us whether to fill the entries with links, or
|
||
FAT_CLUSTER_AVAILABLE
|
||
|
||
|
||
Return Value:
|
||
|
||
VOID
|
||
|
||
--*/
|
||
|
||
{
|
||
#define MAXCOUNTCLUS 0x10000
|
||
#define COUNTSAVEDBCBS ((MAXCOUNTCLUS * sizeof(FAT_ENTRY) / PAGE_SIZE) + 2)
|
||
PBCB SavedBcbs[COUNTSAVEDBCBS][2];
|
||
|
||
ULONG SectorSize;
|
||
ULONG Cluster;
|
||
|
||
LBO StartSectorLbo;
|
||
LBO FinalSectorLbo;
|
||
LBO Lbo;
|
||
|
||
PVOID PinnedFat;
|
||
|
||
BOOLEAN ReleaseMutex = FALSE;
|
||
|
||
ULONG SavedStartingFatIndex = StartingFatIndex;
|
||
|
||
PAGED_CODE();
|
||
|
||
DebugTrace(+1, Dbg, "FatSetFatRun\n", 0);
|
||
DebugTrace( 0, Dbg, " Vcb = %p\n", Vcb);
|
||
DebugTrace( 0, Dbg, " StartingFatIndex = %8x\n", StartingFatIndex);
|
||
DebugTrace( 0, Dbg, " ClusterCount = %8lx\n", ClusterCount);
|
||
DebugTrace( 0, Dbg, " ChainTogether = %s\n", ChainTogether ? "TRUE":"FALSE");
|
||
|
||
//
|
||
// Make sure they gave us a valid fat run.
|
||
//
|
||
|
||
FatVerifyIndexIsValid(IrpContext, Vcb, StartingFatIndex);
|
||
FatVerifyIndexIsValid(IrpContext, Vcb, StartingFatIndex + ClusterCount - 1);
|
||
|
||
//
|
||
// Check special case
|
||
//
|
||
|
||
if (ClusterCount == 0) {
|
||
|
||
DebugTrace(-1, Dbg, "FatSetFatRun -> (VOID)\n", 0);
|
||
return;
|
||
}
|
||
|
||
//
|
||
// Set Sector Size
|
||
//
|
||
|
||
SectorSize = 1 << Vcb->AllocationSupport.LogOfBytesPerSector;
|
||
|
||
//
|
||
// Case on 12 or 16 bit fats.
|
||
//
|
||
// In the 12 bit case (mostly floppies) we always have the whole fat
|
||
// (max 6k bytes) pinned during allocation operations. This is possibly
|
||
// a wee bit slower, but saves headaches over fat entries with 8 bits
|
||
// on one page, and 4 bits on the next.
|
||
//
|
||
// In the 16 bit case we only read one page at a time, as needed.
|
||
//
|
||
|
||
//
|
||
// DEAL WITH 12 BIT CASE
|
||
//
|
||
|
||
_SEH2_TRY {
|
||
|
||
if (Vcb->AllocationSupport.FatIndexBitSize == 12) {
|
||
|
||
//
|
||
// We read in the entire fat. Note that using prepare write marks
|
||
// the bcb pre-dirty, so we don't have to do it explicitly.
|
||
//
|
||
|
||
RtlZeroMemory( &SavedBcbs[0][0], 2 * sizeof(PBCB) * 2);
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
FatReservedBytes( &Vcb->Bpb ),
|
||
FatBytesPerFat( &Vcb->Bpb ),
|
||
&SavedBcbs[0][0],
|
||
&PinnedFat,
|
||
TRUE,
|
||
FALSE );
|
||
|
||
//
|
||
// Mark the affected sectors dirty. Note that FinalSectorLbo is
|
||
// the Lbo of the END of the entry (Thus * 3 + 2). This makes sure
|
||
// we catch the case of a dirty fat entry straddling a sector boundry.
|
||
//
|
||
// Note that if the first AddMcbEntry succeeds, all following ones
|
||
// will simply coalese, and thus also succeed.
|
||
//
|
||
|
||
StartSectorLbo = (FatReservedBytes( &Vcb->Bpb ) + StartingFatIndex * 3 / 2)
|
||
& ~(SectorSize - 1);
|
||
|
||
FinalSectorLbo = (FatReservedBytes( &Vcb->Bpb ) + ((StartingFatIndex +
|
||
ClusterCount) * 3 + 2) / 2) & ~(SectorSize - 1);
|
||
|
||
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
|
||
}
|
||
|
||
//
|
||
// Store the entries into the fat; we need a little
|
||
// synchonization here and can't use a spinlock since the bytes
|
||
// might not be resident.
|
||
//
|
||
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
|
||
for (Cluster = StartingFatIndex;
|
||
Cluster < StartingFatIndex + ClusterCount - 1;
|
||
Cluster++) {
|
||
|
||
FatSet12BitEntry( PinnedFat,
|
||
Cluster,
|
||
ChainTogether ? Cluster + 1 : FAT_CLUSTER_AVAILABLE );
|
||
}
|
||
|
||
//
|
||
// Save the last entry
|
||
//
|
||
|
||
FatSet12BitEntry( PinnedFat,
|
||
Cluster,
|
||
ChainTogether ?
|
||
FAT_CLUSTER_LAST & 0xfff : FAT_CLUSTER_AVAILABLE );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
|
||
} else if (Vcb->AllocationSupport.FatIndexBitSize == 32) {
|
||
|
||
//
|
||
// DEAL WITH 32 BIT CASE
|
||
//
|
||
|
||
for (;;) {
|
||
|
||
VBO StartOffsetInVolume;
|
||
VBO FinalOffsetInVolume;
|
||
|
||
ULONG Page;
|
||
ULONG FinalCluster;
|
||
PULONG FatEntry = NULL;
|
||
ULONG ClusterCountThisRun;
|
||
|
||
StartOffsetInVolume = FatReservedBytes(&Vcb->Bpb) +
|
||
StartingFatIndex * sizeof(FAT_ENTRY);
|
||
|
||
if (ClusterCount > MAXCOUNTCLUS) {
|
||
ClusterCountThisRun = MAXCOUNTCLUS;
|
||
} else {
|
||
ClusterCountThisRun = ClusterCount;
|
||
}
|
||
|
||
FinalOffsetInVolume = StartOffsetInVolume +
|
||
(ClusterCountThisRun - 1) * sizeof(FAT_ENTRY);
|
||
|
||
{
|
||
ULONG NumberOfPages;
|
||
ULONG Offset;
|
||
|
||
NumberOfPages = (FinalOffsetInVolume / PAGE_SIZE) -
|
||
(StartOffsetInVolume / PAGE_SIZE) + 1;
|
||
|
||
RtlZeroMemory( &SavedBcbs[0][0], (NumberOfPages + 1) * sizeof(PBCB) * 2 );
|
||
|
||
for ( Page = 0, Offset = StartOffsetInVolume & ~(PAGE_SIZE - 1);
|
||
Page < NumberOfPages;
|
||
Page++, Offset += PAGE_SIZE ) {
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
Offset,
|
||
PAGE_SIZE,
|
||
&SavedBcbs[Page][0],
|
||
(PVOID *)&SavedBcbs[Page][1],
|
||
TRUE,
|
||
FALSE );
|
||
|
||
if (Page == 0) {
|
||
|
||
FatEntry = (PULONG)((PUCHAR)SavedBcbs[0][1] +
|
||
(StartOffsetInVolume % PAGE_SIZE));
|
||
}
|
||
}
|
||
}
|
||
|
||
//
|
||
// Mark the run dirty
|
||
//
|
||
|
||
StartSectorLbo = StartOffsetInVolume & ~(SectorSize - 1);
|
||
FinalSectorLbo = FinalOffsetInVolume & ~(SectorSize - 1);
|
||
|
||
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO)Lbo, Lbo, SectorSize );
|
||
}
|
||
|
||
//
|
||
// Store the entries
|
||
//
|
||
// We need extra synchronization here for broken architectures
|
||
// like the ALPHA that don't support atomic 16 bit writes.
|
||
//
|
||
|
||
#ifdef ALPHA
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
#endif // ALPHA
|
||
|
||
FinalCluster = StartingFatIndex + ClusterCountThisRun - 1;
|
||
Page = 0;
|
||
|
||
for (Cluster = StartingFatIndex;
|
||
Cluster <= FinalCluster;
|
||
Cluster++, FatEntry++) {
|
||
|
||
//
|
||
// If we just crossed a page boundry (as opposed to starting
|
||
// on one), update our idea of FatEntry.
|
||
|
||
if ( (((ULONG_PTR)FatEntry & (PAGE_SIZE-1)) == 0) &&
|
||
(Cluster != StartingFatIndex) ) {
|
||
|
||
Page += 1;
|
||
FatEntry = (PULONG)SavedBcbs[Page][1];
|
||
}
|
||
|
||
*FatEntry = ChainTogether ? (FAT_ENTRY)(Cluster + 1) :
|
||
FAT_CLUSTER_AVAILABLE;
|
||
}
|
||
|
||
//
|
||
// Fix up the last entry if we were chaining together
|
||
//
|
||
|
||
if ((ClusterCount <= MAXCOUNTCLUS) &&
|
||
ChainTogether ) {
|
||
|
||
*(FatEntry-1) = FAT_CLUSTER_LAST;
|
||
}
|
||
|
||
#ifdef ALPHA
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
#endif // ALPHA
|
||
|
||
{
|
||
ULONG i;
|
||
|
||
//
|
||
// Unpin the Bcbs
|
||
//
|
||
|
||
for (i = 0; (i < COUNTSAVEDBCBS) && (SavedBcbs[i][0] != NULL); i++) {
|
||
|
||
FatUnpinBcb( IrpContext, SavedBcbs[i][0] );
|
||
SavedBcbs[i][0] = NULL;
|
||
}
|
||
}
|
||
|
||
if (ClusterCount <= MAXCOUNTCLUS) {
|
||
|
||
break;
|
||
|
||
} else {
|
||
|
||
StartingFatIndex += MAXCOUNTCLUS;
|
||
ClusterCount -= MAXCOUNTCLUS;
|
||
}
|
||
}
|
||
|
||
} else {
|
||
|
||
//
|
||
// DEAL WITH 16 BIT CASE
|
||
//
|
||
|
||
VBO StartOffsetInVolume;
|
||
VBO FinalOffsetInVolume;
|
||
|
||
ULONG Page;
|
||
ULONG FinalCluster;
|
||
PUSHORT FatEntry = NULL;
|
||
|
||
StartOffsetInVolume = FatReservedBytes(&Vcb->Bpb) +
|
||
StartingFatIndex * sizeof(USHORT);
|
||
|
||
FinalOffsetInVolume = StartOffsetInVolume +
|
||
(ClusterCount - 1) * sizeof(USHORT);
|
||
|
||
//
|
||
// Read in one page of fat at a time. We cannot read in the
|
||
// all of the fat we need because of cache manager limitations.
|
||
//
|
||
// SavedBcb was initialized to be able to hold the largest
|
||
// possible number of pages in a fat plus and extra one to
|
||
// accomadate the boot sector, plus one more to make sure there
|
||
// is enough room for the RtlZeroMemory below that needs the mark
|
||
// the first Bcb after all the ones we will use as an end marker.
|
||
//
|
||
|
||
{
|
||
ULONG NumberOfPages;
|
||
ULONG Offset;
|
||
|
||
NumberOfPages = (FinalOffsetInVolume / PAGE_SIZE) -
|
||
(StartOffsetInVolume / PAGE_SIZE) + 1;
|
||
|
||
RtlZeroMemory( &SavedBcbs[0][0], (NumberOfPages + 1) * sizeof(PBCB) * 2 );
|
||
|
||
for ( Page = 0, Offset = StartOffsetInVolume & ~(PAGE_SIZE - 1);
|
||
Page < NumberOfPages;
|
||
Page++, Offset += PAGE_SIZE ) {
|
||
|
||
FatPrepareWriteVolumeFile( IrpContext,
|
||
Vcb,
|
||
Offset,
|
||
PAGE_SIZE,
|
||
&SavedBcbs[Page][0],
|
||
(PVOID *)&SavedBcbs[Page][1],
|
||
TRUE,
|
||
FALSE );
|
||
|
||
if (Page == 0) {
|
||
|
||
FatEntry = (PUSHORT)((PUCHAR)SavedBcbs[0][1] +
|
||
(StartOffsetInVolume % PAGE_SIZE));
|
||
}
|
||
}
|
||
}
|
||
|
||
//
|
||
// Mark the run dirty
|
||
//
|
||
|
||
StartSectorLbo = StartOffsetInVolume & ~(SectorSize - 1);
|
||
FinalSectorLbo = FinalOffsetInVolume & ~(SectorSize - 1);
|
||
|
||
for (Lbo = StartSectorLbo; Lbo <= FinalSectorLbo; Lbo += SectorSize) {
|
||
|
||
FatAddMcbEntry( Vcb, &Vcb->DirtyFatMcb, (VBO) Lbo, Lbo, SectorSize );
|
||
}
|
||
|
||
//
|
||
// Store the entries
|
||
//
|
||
// We need extra synchronization here for broken architectures
|
||
// like the ALPHA that don't support atomic 16 bit writes.
|
||
//
|
||
|
||
#ifdef ALPHA
|
||
FatLockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = TRUE;
|
||
#endif // ALPHA
|
||
|
||
FinalCluster = StartingFatIndex + ClusterCount - 1;
|
||
Page = 0;
|
||
|
||
for (Cluster = StartingFatIndex;
|
||
Cluster <= FinalCluster;
|
||
Cluster++, FatEntry++) {
|
||
|
||
//
|
||
// If we just crossed a page boundry (as opposed to starting
|
||
// on one), update our idea of FatEntry.
|
||
|
||
if ( (((ULONG_PTR)FatEntry & (PAGE_SIZE-1)) == 0) &&
|
||
(Cluster != StartingFatIndex) ) {
|
||
|
||
Page += 1;
|
||
FatEntry = (PUSHORT)SavedBcbs[Page][1];
|
||
}
|
||
|
||
*FatEntry = (USHORT) (ChainTogether ? (FAT_ENTRY)(Cluster + 1) :
|
||
FAT_CLUSTER_AVAILABLE);
|
||
}
|
||
|
||
//
|
||
// Fix up the last entry if we were chaining together
|
||
//
|
||
|
||
if ( ChainTogether ) {
|
||
|
||
#ifdef _MSC_VER
|
||
#pragma warning( suppress: 4310 )
|
||
#endif
|
||
*(FatEntry-1) = (USHORT)FAT_CLUSTER_LAST;
|
||
|
||
}
|
||
#ifdef ALPHA
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
ReleaseMutex = FALSE;
|
||
#endif // ALPHA
|
||
}
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
ULONG i;
|
||
|
||
DebugUnwind( FatSetFatRun );
|
||
|
||
//
|
||
// If we still somehow have the Mutex, release it.
|
||
//
|
||
|
||
if (ReleaseMutex) {
|
||
|
||
NT_ASSERT( _SEH2_AbnormalTermination() );
|
||
|
||
FatUnlockFreeClusterBitMap( Vcb );
|
||
}
|
||
|
||
//
|
||
// Unpin the Bcbs
|
||
//
|
||
|
||
for (i = 0; (i < COUNTSAVEDBCBS) && (SavedBcbs[i][0] != NULL); i++) {
|
||
|
||
FatUnpinBcb( IrpContext, SavedBcbs[i][0] );
|
||
}
|
||
|
||
//
|
||
// At this point nothing in this finally clause should have raised.
|
||
// So, now comes the unsafe (sigh) stuff.
|
||
//
|
||
|
||
if ( _SEH2_AbnormalTermination() &&
|
||
(Vcb->AllocationSupport.FatIndexBitSize == 32) ) {
|
||
|
||
//
|
||
// Fat32 unwind
|
||
//
|
||
// This case is more complex because the FAT12 and FAT16 cases
|
||
// pin all the needed FAT pages (128K max), after which it
|
||
// can't fail, before changing any FAT entries. In the Fat32
|
||
// case, it may not be practical to pin all the needed FAT
|
||
// pages, because that could span many megabytes. So Fat32
|
||
// attacks in chunks, and if a failure occurs once the first
|
||
// chunk has been updated, we have to back out the updates.
|
||
//
|
||
// The unwind consists of walking back over each FAT entry we
|
||
// have changed, setting it back to the previous value. Note
|
||
// that the previous value with either be FAT_CLUSTER_AVAILABLE
|
||
// (if ChainTogether==TRUE) or a simple link to the successor
|
||
// (if ChainTogether==FALSE).
|
||
//
|
||
// We concede that any one of these calls could fail too; our
|
||
// objective is to make this case no more likely than the case
|
||
// for a file consisting of multiple disjoint runs.
|
||
//
|
||
|
||
while ( StartingFatIndex > SavedStartingFatIndex ) {
|
||
|
||
StartingFatIndex--;
|
||
|
||
FatSetFatEntry( IrpContext, Vcb, StartingFatIndex,
|
||
ChainTogether ?
|
||
StartingFatIndex + 1 : FAT_CLUSTER_AVAILABLE );
|
||
}
|
||
}
|
||
|
||
DebugTrace(-1, Dbg, "FatSetFatRun -> (VOID)\n", 0);
|
||
} _SEH2_END;
|
||
|
||
return;
|
||
}
|
||
|
||
|
||
//
|
||
// Internal support routine
|
||
//
|
||
|
||
UCHAR
|
||
FatLogOf (
|
||
IN ULONG Value
|
||
)
|
||
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine just computes the base 2 log of an integer. It is only used
|
||
on objects that are know to be powers of two.
|
||
|
||
Arguments:
|
||
|
||
Value - The value to take the base 2 log of.
|
||
|
||
Return Value:
|
||
|
||
UCHAR - The base 2 log of Value.
|
||
|
||
--*/
|
||
|
||
{
|
||
UCHAR Log = 0;
|
||
|
||
#if FASTFATDBG
|
||
ULONG OrigValue = Value;
|
||
#endif
|
||
|
||
PAGED_CODE();
|
||
|
||
//
|
||
// Knock bits off until we we get a one at position 0
|
||
//
|
||
|
||
while ( (Value & 0xfffffffe) != 0 ) {
|
||
|
||
Log++;
|
||
Value >>= 1;
|
||
}
|
||
|
||
//
|
||
// If there was more than one bit set, the file system messed up,
|
||
// Bug Check.
|
||
//
|
||
|
||
if (Value != 0x1) {
|
||
|
||
DebugTrace(+1, Dbg, "LogOf\n", 0);
|
||
DebugTrace( 0, Dbg, " Value = %8lx\n", OrigValue);
|
||
|
||
DebugTrace( 0, Dbg, "Received non power of 2.\n", 0);
|
||
|
||
DebugTrace(-1, Dbg, "LogOf -> %8lx\n", Log);
|
||
|
||
#ifdef _MSC_VER
|
||
#pragma prefast( suppress: 28159, "we bugcheck here because our internal data structures are seriously corrupted if this happens" )
|
||
#endif
|
||
FatBugCheck( Value, Log, 0 );
|
||
}
|
||
|
||
return Log;
|
||
}
|
||
|
||
|
||
VOID
|
||
FatExamineFatEntries(
|
||
IN PIRP_CONTEXT IrpContext,
|
||
IN PVCB Vcb,
|
||
IN ULONG StartIndex OPTIONAL,
|
||
IN ULONG EndIndex OPTIONAL,
|
||
IN BOOLEAN SetupWindows,
|
||
IN PFAT_WINDOW SwitchToWindow OPTIONAL,
|
||
IN PULONG BitMapBuffer OPTIONAL
|
||
)
|
||
/*++
|
||
|
||
Routine Description:
|
||
|
||
This routine handles scanning a segment of the FAT into in-memory structures.
|
||
|
||
There are three fundamental cases, with variations depending on the FAT type:
|
||
|
||
1) During volume setup, FatSetupAllocations
|
||
|
||
1a) for FAT12/16, read the FAT into our free clusterbitmap
|
||
1b) for FAT32, perform the initial scan for window free cluster counts
|
||
|
||
2) Switching FAT32 windows on the fly during system operation
|
||
|
||
3) Reading arbitrary segments of the FAT for the purposes of the GetVolumeBitmap
|
||
call (only for FAT32)
|
||
|
||
There really is too much going on in here. At some point this should be
|
||
substantially rewritten.
|
||
|
||
Arguments:
|
||
|
||
Vcb - Supplies the volume involved
|
||
|
||
StartIndex - Supplies the starting cluster, ignored if SwitchToWindow supplied
|
||
|
||
EndIndex - Supplies the ending cluster, ignored if SwitchToWindow supplied
|
||
|
||
SetupWindows - Indicates if we are doing the initial FAT32 scan
|
||
|
||
SwitchToWindow - Supplies the FAT window we are examining and will switch to
|
||
|
||
BitMapBuffer - Supplies a specific bitmap to fill in, if not supplied we fill
|
||
in the volume free cluster bitmap if !SetupWindows
|
||
|
||
Return Value:
|
||
|
||
None. Lots of side effects.
|
||
|
||
--*/
|
||
{
|
||
ULONG FatIndexBitSize;
|
||
ULONG Page = 0;
|
||
ULONG Offset = 0;
|
||
ULONG FatIndex;
|
||
FAT_ENTRY FatEntry = FAT_CLUSTER_AVAILABLE;
|
||
FAT_ENTRY FirstFatEntry = FAT_CLUSTER_AVAILABLE;
|
||
PUSHORT FatBuffer;
|
||
PVOID pv;
|
||
PBCB Bcb = NULL;
|
||
ULONG EntriesPerWindow;
|
||
|
||
ULONG ClustersThisRun;
|
||
ULONG StartIndexOfThisRun;
|
||
|
||
PULONG FreeClusterCount = NULL;
|
||
|
||
PFAT_WINDOW CurrentWindow = NULL;
|
||
|
||
PVOID NewBitMapBuffer = NULL;
|
||
PRTL_BITMAP BitMap = NULL;
|
||
RTL_BITMAP PrivateBitMap;
|
||
|
||
ULONG ClusterSize = 0;
|
||
ULONG PrefetchPages = 0;
|
||
ULONG FatPages = 0;
|
||
|
||
VBO BadClusterVbo = 0;
|
||
LBO Lbo = 0;
|
||
|
||
enum RunType {
|
||
FreeClusters,
|
||
AllocatedClusters,
|
||
UnknownClusters
|
||
} CurrentRun;
|
||
|
||
PAGED_CODE();
|
||
|
||
//
|
||
// Now assert correct usage.
|
||
//
|
||
|
||
FatIndexBitSize = Vcb->AllocationSupport.FatIndexBitSize;
|
||
|
||
NT_ASSERT( !(SetupWindows && (SwitchToWindow || BitMapBuffer)));
|
||
NT_ASSERT( !(SetupWindows && FatIndexBitSize != 32));
|
||
|
||
if (Vcb->NumberOfWindows > 1) {
|
||
|
||
//
|
||
// FAT32: Calculate the number of FAT entries covered by a window. This is
|
||
// equal to the number of bits in the freespace bitmap, the size of which
|
||
// is hardcoded.
|
||
//
|
||
|
||
EntriesPerWindow = MAX_CLUSTER_BITMAP_SIZE;
|
||
|
||
} else {
|
||
|
||
EntriesPerWindow = Vcb->AllocationSupport.NumberOfClusters;
|
||
}
|
||
|
||
//
|
||
// We will also fill in the cumulative count of free clusters for
|
||
// the entire volume. If this is not appropriate, NULL it out
|
||
// shortly.
|
||
//
|
||
|
||
FreeClusterCount = &Vcb->AllocationSupport.NumberOfFreeClusters;
|
||
|
||
if (SetupWindows) {
|
||
|
||
NT_ASSERT(BitMapBuffer == NULL);
|
||
|
||
//
|
||
// In this case we're just supposed to scan the fat and set up
|
||
// the information regarding where the buckets fall and how many
|
||
// free clusters are in each.
|
||
//
|
||
// It is fine to monkey with the real windows, we must be able
|
||
// to do this to activate the volume.
|
||
//
|
||
|
||
BitMap = NULL;
|
||
|
||
CurrentWindow = &Vcb->Windows[0];
|
||
CurrentWindow->FirstCluster = StartIndex;
|
||
CurrentWindow->ClustersFree = 0;
|
||
|
||
//
|
||
// We always wish to calculate total free clusters when
|
||
// setting up the FAT windows.
|
||
//
|
||
|
||
} else if (BitMapBuffer == NULL) {
|
||
|
||
//
|
||
// We will be filling in the free cluster bitmap for the volume.
|
||
// Careful, we can raise out of here and be hopelessly hosed if
|
||
// we built this up in the main bitmap/window itself.
|
||
//
|
||
// For simplicity's sake, we'll do the swap for everyone. FAT32
|
||
// provokes the need since we can't tolerate partial results
|
||
// when switching windows.
|
||
//
|
||
|
||
NT_ASSERT( SwitchToWindow );
|
||
|
||
CurrentWindow = SwitchToWindow;
|
||
StartIndex = CurrentWindow->FirstCluster;
|
||
EndIndex = CurrentWindow->LastCluster;
|
||
|
||
BitMap = &PrivateBitMap;
|
||
NewBitMapBuffer = FsRtlAllocatePoolWithTag( PagedPool,
|
||
(EntriesPerWindow + 7) / 8,
|
||
TAG_FAT_BITMAP );
|
||
|
||
RtlInitializeBitMap( &PrivateBitMap,
|
||
NewBitMapBuffer,
|
||
EndIndex - StartIndex + 1);
|
||
|
||
if ((FatIndexBitSize == 32) &&
|
||
(Vcb->NumberOfWindows > 1)) {
|
||
|
||
//
|
||
// We do not wish count total clusters here.
|
||
//
|
||
|
||
FreeClusterCount = NULL;
|
||
|
||
}
|
||
|
||
} else {
|
||
|
||
BitMap = &PrivateBitMap;
|
||
RtlInitializeBitMap(&PrivateBitMap,
|
||
BitMapBuffer,
|
||
EndIndex - StartIndex + 1);
|
||
|
||
//
|
||
// We do not count total clusters here.
|
||
//
|
||
|
||
FreeClusterCount = NULL;
|
||
}
|
||
|
||
//
|
||
// Now, our start index better be in the file heap.
|
||
//
|
||
|
||
NT_ASSERT( StartIndex >= 2 );
|
||
|
||
_SEH2_TRY {
|
||
|
||
//
|
||
// Pick up the initial chunk of the FAT and first entry.
|
||
//
|
||
|
||
if (FatIndexBitSize == 12) {
|
||
|
||
//
|
||
// We read in the entire fat in the 12 bit case.
|
||
//
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
FatReservedBytes( &Vcb->Bpb ),
|
||
FatBytesPerFat( &Vcb->Bpb ),
|
||
&Bcb,
|
||
(PVOID *)&FatBuffer );
|
||
|
||
FatLookup12BitEntry(FatBuffer, 0, &FirstFatEntry);
|
||
|
||
} else {
|
||
|
||
//
|
||
// Read in one page of fat at a time. We cannot read in the
|
||
// all of the fat we need because of cache manager limitations.
|
||
//
|
||
|
||
ULONG BytesPerEntry = FatIndexBitSize >> 3;
|
||
|
||
FatPages = (FatReservedBytes(&Vcb->Bpb) + FatBytesPerFat(&Vcb->Bpb) + (PAGE_SIZE - 1)) / PAGE_SIZE;
|
||
Page = (FatReservedBytes(&Vcb->Bpb) + StartIndex * BytesPerEntry) / PAGE_SIZE;
|
||
|
||
Offset = Page * PAGE_SIZE;
|
||
|
||
//
|
||
// Prefetch the FAT entries in memory for optimal performance.
|
||
//
|
||
|
||
PrefetchPages = FatPages - Page;
|
||
|
||
if (PrefetchPages > FAT_PREFETCH_PAGE_COUNT) {
|
||
|
||
PrefetchPages = ALIGN_UP_BY(Page, FAT_PREFETCH_PAGE_COUNT) - Page;
|
||
}
|
||
|
||
#if (NTDDI_VERSION >= NTDDI_WIN8)
|
||
FatPrefetchPages( IrpContext,
|
||
Vcb->VirtualVolumeFile,
|
||
Page,
|
||
PrefetchPages );
|
||
#endif
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
Offset,
|
||
PAGE_SIZE,
|
||
&Bcb,
|
||
&pv);
|
||
|
||
if (FatIndexBitSize == 32) {
|
||
|
||
FatBuffer = (PUSHORT)((PUCHAR)pv +
|
||
(FatReservedBytes(&Vcb->Bpb) + StartIndex * BytesPerEntry) %
|
||
PAGE_SIZE);
|
||
|
||
FirstFatEntry = *((PULONG)FatBuffer);
|
||
FirstFatEntry = FirstFatEntry & FAT32_ENTRY_MASK;
|
||
|
||
} else {
|
||
|
||
FatBuffer = (PUSHORT)((PUCHAR)pv +
|
||
FatReservedBytes(&Vcb->Bpb) % PAGE_SIZE) + 2;
|
||
|
||
FirstFatEntry = *FatBuffer;
|
||
}
|
||
|
||
}
|
||
|
||
ClusterSize = 1 << (Vcb->AllocationSupport.LogOfBytesPerCluster);
|
||
|
||
CurrentRun = (FirstFatEntry == FAT_CLUSTER_AVAILABLE) ?
|
||
FreeClusters : AllocatedClusters;
|
||
|
||
StartIndexOfThisRun = StartIndex;
|
||
|
||
for (FatIndex = StartIndex; FatIndex <= EndIndex; FatIndex++) {
|
||
|
||
if (FatIndexBitSize == 12) {
|
||
|
||
FatLookup12BitEntry(FatBuffer, FatIndex, &FatEntry);
|
||
|
||
} else {
|
||
|
||
//
|
||
// If we are setting up the FAT32 windows and have stepped into a new
|
||
// bucket, finalize this one and move forward.
|
||
//
|
||
|
||
if (SetupWindows &&
|
||
FatIndex > StartIndex &&
|
||
(FatIndex - 2) % EntriesPerWindow == 0) {
|
||
|
||
CurrentWindow->LastCluster = FatIndex - 1;
|
||
|
||
if (CurrentRun == FreeClusters) {
|
||
|
||
//
|
||
// We must be counting clusters in order to modify the
|
||
// contents of the window.
|
||
//
|
||
|
||
NT_ASSERT( FreeClusterCount );
|
||
|
||
ClustersThisRun = FatIndex - StartIndexOfThisRun;
|
||
CurrentWindow->ClustersFree += ClustersThisRun;
|
||
|
||
if (FreeClusterCount) {
|
||
*FreeClusterCount += ClustersThisRun;
|
||
}
|
||
|
||
} else {
|
||
|
||
NT_ASSERT(CurrentRun == AllocatedClusters);
|
||
|
||
}
|
||
|
||
StartIndexOfThisRun = FatIndex;
|
||
CurrentRun = UnknownClusters;
|
||
|
||
CurrentWindow++;
|
||
CurrentWindow->ClustersFree = 0;
|
||
CurrentWindow->FirstCluster = FatIndex;
|
||
}
|
||
|
||
//
|
||
// If we just stepped onto a new page, grab a new pointer.
|
||
//
|
||
|
||
if (((ULONG_PTR)FatBuffer & (PAGE_SIZE - 1)) == 0) {
|
||
|
||
FatUnpinBcb( IrpContext, Bcb );
|
||
|
||
Page++;
|
||
Offset += PAGE_SIZE;
|
||
|
||
#if (NTDDI_VERSION >= NTDDI_WIN8)
|
||
//
|
||
// If we have exhausted all the prefetch pages, prefetch the next chunk.
|
||
//
|
||
|
||
if (--PrefetchPages == 0) {
|
||
|
||
PrefetchPages = FatPages - Page;
|
||
|
||
if (PrefetchPages > FAT_PREFETCH_PAGE_COUNT) {
|
||
|
||
PrefetchPages = FAT_PREFETCH_PAGE_COUNT;
|
||
}
|
||
|
||
FatPrefetchPages( IrpContext,
|
||
Vcb->VirtualVolumeFile,
|
||
Page,
|
||
PrefetchPages );
|
||
}
|
||
#endif
|
||
|
||
FatReadVolumeFile( IrpContext,
|
||
Vcb,
|
||
Offset,
|
||
PAGE_SIZE,
|
||
&Bcb,
|
||
&pv );
|
||
|
||
FatBuffer = (PUSHORT)pv;
|
||
}
|
||
|
||
if (FatIndexBitSize == 32) {
|
||
|
||
#ifndef __REACTOS__
|
||
#ifdef _MSC_VER
|
||
#pragma warning( suppress: 4213 )
|
||
#endif
|
||
FatEntry = *((PULONG)FatBuffer)++;
|
||
FatEntry = FatEntry & FAT32_ENTRY_MASK;
|
||
#else
|
||
FatEntry = *FatBuffer;
|
||
FatBuffer += 1;
|
||
FatEntry = FatEntry & FAT32_ENTRY_MASK;
|
||
#endif
|
||
|
||
} else {
|
||
|
||
FatEntry = *FatBuffer;
|
||
FatBuffer += 1;
|
||
}
|
||
}
|
||
|
||
if (CurrentRun == UnknownClusters) {
|
||
|
||
CurrentRun = (FatEntry == FAT_CLUSTER_AVAILABLE) ?
|
||
FreeClusters : AllocatedClusters;
|
||
}
|
||
|
||
//
|
||
// Are we switching from a free run to an allocated run?
|
||
//
|
||
|
||
if (CurrentRun == FreeClusters &&
|
||
FatEntry != FAT_CLUSTER_AVAILABLE) {
|
||
|
||
ClustersThisRun = FatIndex - StartIndexOfThisRun;
|
||
|
||
if (FreeClusterCount) {
|
||
|
||
*FreeClusterCount += ClustersThisRun;
|
||
CurrentWindow->ClustersFree += ClustersThisRun;
|
||
}
|
||
|
||
if (BitMap) {
|
||
|
||
RtlClearBits( BitMap,
|
||
StartIndexOfThisRun - StartIndex,
|
||
ClustersThisRun );
|
||
}
|
||
|
||
CurrentRun = AllocatedClusters;
|
||
StartIndexOfThisRun = FatIndex;
|
||
}
|
||
|
||
//
|
||
// Are we switching from an allocated run to a free run?
|
||
//
|
||
|
||
if (CurrentRun == AllocatedClusters &&
|
||
FatEntry == FAT_CLUSTER_AVAILABLE) {
|
||
|
||
ClustersThisRun = FatIndex - StartIndexOfThisRun;
|
||
|
||
if (BitMap) {
|
||
|
||
RtlSetBits( BitMap,
|
||
StartIndexOfThisRun - StartIndex,
|
||
ClustersThisRun );
|
||
}
|
||
|
||
CurrentRun = FreeClusters;
|
||
StartIndexOfThisRun = FatIndex;
|
||
}
|
||
|
||
//
|
||
// If the entry is marked bad, add it to the bad block MCB
|
||
//
|
||
|
||
if ((SetupWindows || (Vcb->NumberOfWindows == 1)) &&
|
||
(FatInterpretClusterType( Vcb, FatEntry ) == FatClusterBad)) {
|
||
|
||
//
|
||
// This cluster is marked bad.
|
||
// Add it to the BadBlockMcb.
|
||
//
|
||
|
||
Lbo = FatGetLboFromIndex( Vcb, FatIndex );
|
||
FatAddMcbEntry( Vcb, &Vcb->BadBlockMcb, BadClusterVbo, Lbo, ClusterSize );
|
||
BadClusterVbo += ClusterSize;
|
||
}
|
||
}
|
||
|
||
//
|
||
// If we finished the scan, then we know about all the possible bad clusters.
|
||
//
|
||
|
||
SetFlag( Vcb->VcbState, VCB_STATE_FLAG_BAD_BLOCKS_POPULATED);
|
||
|
||
//
|
||
// Now we have to record the final run we encountered
|
||
//
|
||
|
||
ClustersThisRun = FatIndex - StartIndexOfThisRun;
|
||
|
||
if (CurrentRun == FreeClusters) {
|
||
|
||
if (FreeClusterCount) {
|
||
|
||
*FreeClusterCount += ClustersThisRun;
|
||
CurrentWindow->ClustersFree += ClustersThisRun;
|
||
}
|
||
|
||
if (BitMap) {
|
||
|
||
RtlClearBits( BitMap,
|
||
StartIndexOfThisRun - StartIndex,
|
||
ClustersThisRun );
|
||
}
|
||
|
||
} else {
|
||
|
||
if (BitMap) {
|
||
|
||
RtlSetBits( BitMap,
|
||
StartIndexOfThisRun - StartIndex,
|
||
ClustersThisRun );
|
||
}
|
||
}
|
||
|
||
//
|
||
// And finish the last window if we are in setup.
|
||
//
|
||
|
||
if (SetupWindows) {
|
||
|
||
CurrentWindow->LastCluster = FatIndex - 1;
|
||
}
|
||
|
||
//
|
||
// Now switch the active window if required. We've succesfully gotten everything
|
||
// nailed down.
|
||
//
|
||
// If we were tracking the free cluster count, this means we should update the
|
||
// window. This is the case of FAT12/16 initialization.
|
||
//
|
||
|
||
if (SwitchToWindow) {
|
||
|
||
if (Vcb->FreeClusterBitMap.Buffer) {
|
||
|
||
ExFreePool( Vcb->FreeClusterBitMap.Buffer );
|
||
}
|
||
|
||
RtlInitializeBitMap( &Vcb->FreeClusterBitMap,
|
||
NewBitMapBuffer,
|
||
EndIndex - StartIndex + 1 );
|
||
|
||
NewBitMapBuffer = NULL;
|
||
|
||
Vcb->CurrentWindow = SwitchToWindow;
|
||
Vcb->ClusterHint = (ULONG)-1;
|
||
|
||
if (FreeClusterCount) {
|
||
|
||
NT_ASSERT( !SetupWindows );
|
||
|
||
Vcb->CurrentWindow->ClustersFree = *FreeClusterCount;
|
||
}
|
||
}
|
||
|
||
//
|
||
// Make sure plausible things occured ...
|
||
//
|
||
|
||
if (!SetupWindows && BitMapBuffer == NULL) {
|
||
|
||
ASSERT_CURRENT_WINDOW_GOOD( Vcb );
|
||
}
|
||
|
||
NT_ASSERT(Vcb->AllocationSupport.NumberOfFreeClusters <= Vcb->AllocationSupport.NumberOfClusters);
|
||
|
||
} _SEH2_FINALLY {
|
||
|
||
//
|
||
// Unpin the last bcb and drop the temporary bitmap buffer if it exists.
|
||
//
|
||
|
||
FatUnpinBcb( IrpContext, Bcb);
|
||
|
||
if (NewBitMapBuffer) {
|
||
|
||
ExFreePool( NewBitMapBuffer );
|
||
}
|
||
} _SEH2_END;
|
||
}
|
||
|