reactos/drivers/filesystems/fastfat_new/cachesup.c

2018 lines
52 KiB
C
Raw Normal View History

/*++
Copyright (c) 1990-2000 Microsoft Corporation
Module Name:
cache.c
Abstract:
This module implements the cache management routines for the Fat
FSD and FSP, by calling the Common Cache Manager.
--*/
#include "fatprocs.h"
//
// The Bug check file id for this module
//
#define BugCheckFileId (FAT_BUG_CHECK_CACHESUP)
//
// Local debug trace level
//
#define Dbg (DEBUG_TRACE_CACHESUP)
#if DBG
BOOLEAN
FatIsCurrentOperationSynchedForDcbTeardown (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb
);
#endif
#ifdef ALLOC_PRAGMA
#pragma alloc_text(PAGE, FatCloseEaFile)
#pragma alloc_text(PAGE, FatCompleteMdl)
#pragma alloc_text(PAGE, FatOpenDirectoryFile)
#pragma alloc_text(PAGE, FatOpenEaFile)
#pragma alloc_text(PAGE, FatPinMappedData)
#pragma alloc_text(PAGE, FatPrepareWriteDirectoryFile)
#pragma alloc_text(PAGE, FatPrepareWriteVolumeFile)
#pragma alloc_text(PAGE, FatReadDirectoryFile)
#pragma alloc_text(PAGE, FatReadVolumeFile)
#pragma alloc_text(PAGE, FatRepinBcb)
#pragma alloc_text(PAGE, FatSyncUninitializeCacheMap)
#pragma alloc_text(PAGE, FatUnpinRepinnedBcbs)
#pragma alloc_text(PAGE, FatZeroData)
#pragma alloc_text(PAGE, FatPrefetchPages)
#if DBG
#pragma alloc_text(PAGE, FatIsCurrentOperationSynchedForDcbTeardown)
#endif
#endif
VOID
FatInitializeCacheMap (
_In_ PFILE_OBJECT FileObject,
_In_ PCC_FILE_SIZES FileSizes,
_In_ BOOLEAN PinAccess,
_In_ PCACHE_MANAGER_CALLBACKS Callbacks,
_In_ PVOID LazyWriteContext
)
/*++
Routine Description:
Wrapper over CcInitializeCacheMap and CcSetAdditionalCacheAttributesEx to initialize
caching and enable IO accounting on a file.
--*/
{
//
// Initialize caching
//
CcInitializeCacheMap( FileObject,
FileSizes,
PinAccess,
Callbacks,
LazyWriteContext );
#if (NTDDI_VERSION >= NTDDI_WIN8)
//
// Enable Disk IO Accounting for this file
//
if (FatDiskAccountingEnabled) {
CcSetAdditionalCacheAttributesEx( FileObject, CC_ENABLE_DISK_IO_ACCOUNTING );
}
#endif
}
VOID
FatReadVolumeFile (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN VBO StartingVbo,
IN ULONG ByteCount,
OUT PBCB *Bcb,
OUT PVOID *Buffer
)
/*++
Routine Description:
This routine is called when the specified range of sectors is to be
read into the cache. In fat, the volume file only contains the boot
sector, reserved sectors, and the "fat(s)." Thus the volume file is
of fixed size and only extends up to (but not not including) the root
directory entry, and will never move or change size.
The fat volume file is also peculiar in that, since it starts at the
logical beginning of the disk, Vbo == Lbo.
Arguments:
Vcb - Pointer to the VCB for the volume
StartingVbo - The virtual offset of the first desired byte
ByteCount - Number of bytes desired
Bcb - Returns a pointer to the BCB which is valid until unpinned
Buffer - Returns a pointer to the sectors, which is valid until unpinned
--*/
{
LARGE_INTEGER Vbo;
PAGED_CODE();
//
// Check to see that all references are within the Bios Parameter Block
// or the fat(s). A special case is made when StartingVbo == 0 at
// mounting time since we do not know how big the fat is.
//
NT_ASSERT( ((StartingVbo == 0) || ((StartingVbo + ByteCount) <= (ULONG)
(FatRootDirectoryLbo( &Vcb->Bpb ) + PAGE_SIZE))));
DebugTrace(+1, Dbg, "FatReadVolumeFile\n", 0);
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
//
// Call the Cache manager to attempt the transfer.
//
Vbo.QuadPart = StartingVbo;
if (!CcMapData( Vcb->VirtualVolumeFile,
&Vbo,
ByteCount,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
Bcb,
Buffer )) {
NT_ASSERT( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
//
// Could not read the data without waiting (cache miss).
//
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
}
DbgDoit( IrpContext->PinCount += 1 )
DebugTrace(-1, Dbg, "FatReadVolumeFile -> VOID, *BCB = %p\n", *Bcb);
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatPrepareWriteVolumeFile (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN VBO StartingVbo,
IN ULONG ByteCount,
OUT PBCB *Bcb,
OUT PVOID *Buffer,
IN BOOLEAN Reversible,
IN BOOLEAN Zero
)
/*++
Routine Description:
This routine first looks to see if the specified range of sectors,
is already in the cache. If so, it increments the BCB PinCount,
sets the BCB dirty, and returns with the location of the sectors.
If the sectors are not in the cache and Wait is TRUE, it finds a
free BCB (potentially causing a flush), and clears out the entire
buffer. Once this is done, it increments the BCB PinCount, sets the
BCB dirty, and returns with the location of the sectors.
If the sectors are not in the cache and Wait is FALSE, this routine
raises STATUS_CANT_WAIT.
Arguments:
Vcb - Pointer to the VCB for the volume
StartingVbo - The virtual offset of the first byte to be written
ByteCount - Number of bytes to be written
Bcb - Returns a pointer to the BCB which is valid until unpinned
Buffer - Returns a pointer to the sectors, which is valid until unpinned
Reversible - Supplies TRUE if the specified range of modification should
be repinned so that the operation can be reversed in a controlled
fashion if errors are encountered.
Zero - Supplies TRUE if the specified range of bytes should be zeroed
--*/
{
LARGE_INTEGER Vbo;
PAGED_CODE();
//
// Check to see that all references are within the Bios Parameter Block
// or the fat(s).
//
NT_ASSERT( ((StartingVbo + ByteCount) <= (ULONG)
(FatRootDirectoryLbo( &Vcb->Bpb ))));
DebugTrace(+1, Dbg, "FatPrepareWriteVolumeFile\n", 0);
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", (ULONG)StartingVbo);
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
DebugTrace( 0, Dbg, "Zero = %08lx\n", Zero);
//
// Call the Cache manager to attempt the transfer.
//
Vbo.QuadPart = StartingVbo;
if (!CcPinRead( Vcb->VirtualVolumeFile,
&Vbo,
ByteCount,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
Bcb,
Buffer )) {
NT_ASSERT( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
//
// Could not read the data without waiting (cache miss).
//
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
}
//
// This keeps the data pinned until we complete the request
// and writes the dirty bit through to the disk.
//
DbgDoit( IrpContext->PinCount += 1 )
_SEH2_TRY {
if (Zero) {
RtlZeroMemory( *Buffer, ByteCount );
}
FatSetDirtyBcb( IrpContext, *Bcb, Vcb, Reversible );
} _SEH2_FINALLY {
if (_SEH2_AbnormalTermination()) {
FatUnpinBcb(IrpContext, *Bcb);
}
} _SEH2_END;
DebugTrace(-1, Dbg, "FatPrepareWriteVolumeFile -> VOID, *Bcb = %p\n", *Bcb);
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatReadDirectoryFile (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb,
IN VBO StartingVbo,
IN ULONG ByteCount,
IN BOOLEAN Pin,
OUT PBCB *Bcb,
OUT PVOID *Buffer,
OUT PNTSTATUS Status
)
/*++
Routine Description:
This routine is called when the specified range of sectors is to be
read into the cache. If the desired range falls beyond the current
cache mapping, the fat will be searched, and if the desired range can
be satisfied, the cache mapping will be extended and the MCB updated
accordingly.
Arguments:
Dcb - Pointer to the DCB for the directory
StartingVbo - The virtual offset of the first desired byte
ByteCount - Number of bytes desired
Pin - Tells us if we should pin instead of just mapping.
Bcb - Returns a pointer to the BCB which is valid until unpinned
Buffer - Returns a pointer to the sectors, which is valid until unpinned
Status - Returns the status of the operation.
--*/
{
LARGE_INTEGER Vbo;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatReadDirectoryFile\n", 0);
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
//
// Check for the zero case
//
if (ByteCount == 0) {
DebugTrace(0, Dbg, "Nothing to read\n", 0);
*Bcb = NULL;
*Buffer = NULL;
*Status = STATUS_SUCCESS;
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID\n", 0);
return;
}
//
// If we need to create a directory file and initialize the
// cachemap, do so.
//
FatOpenDirectoryFile( IrpContext, Dcb );
//
// Now if the transfer is beyond the allocation size return EOF.
//
if (StartingVbo >= Dcb->Header.AllocationSize.LowPart) {
DebugTrace(0, Dbg, "End of file read for directory\n", 0);
*Bcb = NULL;
*Buffer = NULL;
*Status = STATUS_END_OF_FILE;
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID\n", 0);
return;
}
//
// If the caller is trying to read past the EOF, truncate the
// read.
//
ByteCount = (Dcb->Header.AllocationSize.LowPart - StartingVbo < ByteCount) ?
Dcb->Header.AllocationSize.LowPart - StartingVbo : ByteCount;
NT_ASSERT( ByteCount != 0 );
//
// Call the Cache manager to attempt the transfer.
//
Vbo.QuadPart = StartingVbo;
if (Pin ?
!CcPinRead( Dcb->Specific.Dcb.DirectoryFile,
&Vbo,
ByteCount,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
Bcb,
Buffer )
:
!CcMapData( Dcb->Specific.Dcb.DirectoryFile,
&Vbo,
ByteCount,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
Bcb,
Buffer ) ) {
//
// Could not read the data without waiting (cache miss).
//
*Bcb = NULL;
*Buffer = NULL;
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
}
DbgDoit( IrpContext->PinCount += 1 )
*Status = STATUS_SUCCESS;
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID, *BCB = %p\n", *Bcb);
return;
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatPrepareWriteDirectoryFile (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb,
IN VBO StartingVbo,
IN ULONG ByteCount,
OUT PBCB *Bcb,
OUT PVOID *Buffer,
IN BOOLEAN Zero,
IN BOOLEAN Reversible,
OUT PNTSTATUS Status
)
/*++
Routine Description:
This routine first looks to see if the specified range of sectors
is already in the cache. If so, it increments the BCB PinCount,
sets the BCB dirty, and returns TRUE with the location of the sectors.
The IrpContext->Flags .. Wait == TRUE/FALSE actions of this routine are identical to
FatPrepareWriteVolumeFile() above.
Arguments:
Dcb - Pointer to the DCB for the directory
StartingVbo - The virtual offset of the first byte to be written
ByteCount - Number of bytes to be written
Bcb - Returns a pointer to the BCB which is valid until unpinned
Buffer - Returns a pointer to the sectors, which is valid until unpinned
Zero - Supplies TRUE if the specified range of bytes should be zeroed
Reversible - Supplies TRUE if the specified range of modification should
be repinned so that the operation can be reversed in a controlled
fashion if errors are encountered.
Status - Returns the status of the operation.
--*/
{
LARGE_INTEGER Vbo;
ULONG InitialAllocation = 0;
BOOLEAN UnwindWeAllocatedDiskSpace = FALSE;
PBCB LocalBcb = NULL;
PVOID LocalBuffer = NULL;
ULONG InitialRequest = ByteCount;
ULONG MappingGranularity = PAGE_SIZE;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatPrepareWriteDirectoryFile\n", 0);
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", (ULONG)StartingVbo);
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
DebugTrace( 0, Dbg, "Zero = %08lx\n", Zero);
*Bcb = NULL;
*Buffer = NULL;
//
// If we need to create a directory file and initialize the
// cachemap, do so.
//
FatOpenDirectoryFile( IrpContext, Dcb );
//
// If the transfer is beyond the allocation size we need to
// extend the directory's allocation. The call to
// AddFileAllocation will raise a condition if
// it runs out of disk space. Note that the root directory
// cannot be extended.
//
Vbo.QuadPart = StartingVbo;
_SEH2_TRY {
if (StartingVbo + ByteCount > Dcb->Header.AllocationSize.LowPart) {
if (NodeType(Dcb) == FAT_NTC_ROOT_DCB &&
!FatIsFat32(Dcb->Vcb)) {
FatRaiseStatus( IrpContext, STATUS_DISK_FULL );
}
DebugTrace(0, Dbg, "Try extending normal directory\n", 0);
InitialAllocation = Dcb->Header.AllocationSize.LowPart;
FatAddFileAllocation( IrpContext,
Dcb,
Dcb->Specific.Dcb.DirectoryFile,
StartingVbo + ByteCount );
UnwindWeAllocatedDiskSpace = TRUE;
//
// Inform the cache manager of the new allocation
//
Dcb->Header.FileSize.LowPart =
Dcb->Header.AllocationSize.LowPart;
CcSetFileSizes( Dcb->Specific.Dcb.DirectoryFile,
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize );
//
// Set up the Bitmap buffer if it is not big enough already
//
FatCheckFreeDirentBitmap( IrpContext, Dcb );
//
// The newly allocated clusters should be zeroed starting at
// the previous allocation size
//
Zero = TRUE;
Vbo.QuadPart = InitialAllocation;
ByteCount = Dcb->Header.AllocationSize.LowPart - InitialAllocation;
}
while (ByteCount > 0) {
ULONG BytesToPin;
LocalBcb = NULL;
//
// We must pin in terms of pages below the boundary of the initial request.
// Once we pass the end of the request, we are free to expand the pin size to
// VACB_MAPPING_GRANULARITY. This will prevent Cc from returning OBCBs
// and hence will prevent bugchecks when we then attempt to repin one, yet
// allow us to be more efficient by pinning in 256KB chunks instead of 4KB pages.
//
if (Vbo.QuadPart > StartingVbo + InitialRequest) {
MappingGranularity = VACB_MAPPING_GRANULARITY;
}
//
// If the first and final byte are both described by the same page, pin
// the entire range. Note we pin in pages to prevent cache manager from
// returning OBCBs, which would result in a bugcheck on CcRepinBcb.
//
if ((Vbo.QuadPart / MappingGranularity) ==
((Vbo.QuadPart + ByteCount - 1) / MappingGranularity)) {
BytesToPin = ByteCount;
} else {
BytesToPin = MappingGranularity -
((ULONG)Vbo.QuadPart & (MappingGranularity - 1));
}
if (!CcPinRead( Dcb->Specific.Dcb.DirectoryFile,
&Vbo,
BytesToPin,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
&LocalBcb,
&LocalBuffer )) {
//
// Could not read the data without waiting (cache miss).
//
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
}
//
// Update our caller with the beginning of their request.
//
if (*Buffer == NULL) {
*Buffer = LocalBuffer;
*Bcb = LocalBcb;
}
DbgDoit( IrpContext->PinCount += 1 )
if (Zero) {
//
// We set this guy dirty right now so that we can raise CANT_WAIT when
// it needs to be done. It'd be beautiful if we could noop the read IO
// since we know we don't care about it.
//
RtlZeroMemory( LocalBuffer, BytesToPin );
CcSetDirtyPinnedData( LocalBcb, NULL );
}
ByteCount -= BytesToPin;
Vbo.QuadPart += BytesToPin;
if (*Bcb != LocalBcb) {
FatRepinBcb( IrpContext, LocalBcb );
FatUnpinBcb( IrpContext, LocalBcb );
}
}
//
// This lets us get the data pinned until we complete the request
// and writes the dirty bit through to the disk.
//
FatSetDirtyBcb( IrpContext, *Bcb, Dcb->Vcb, Reversible );
*Status = STATUS_SUCCESS;
} _SEH2_FINALLY {
DebugUnwind( FatPrepareWriteDirectoryFile );
if (_SEH2_AbnormalTermination()) {
//
// Make sure we unpin the buffers.
//
if (*Bcb != LocalBcb) {
FatUnpinBcb( IrpContext, LocalBcb );
}
FatUnpinBcb(IrpContext, *Bcb);
//
// These steps are carefully arranged - FatTruncateFileAllocation can raise.
// Make sure we unpin the buffer. If FTFA raises, the effect should be benign.
//
if (UnwindWeAllocatedDiskSpace == TRUE) {
//
// Inform the cache manager of the change.
//
FatTruncateFileAllocation( IrpContext, Dcb, InitialAllocation );
Dcb->Header.FileSize.LowPart =
Dcb->Header.AllocationSize.LowPart;
CcSetFileSizes( Dcb->Specific.Dcb.DirectoryFile,
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize );
}
}
DebugTrace(-1, Dbg, "FatPrepareWriteDirectoryFile -> (VOID), *Bcb = %p\n", *Bcb);
} _SEH2_END;
return;
}
#if DBG
BOOLEAN FatDisableParentCheck = 0;
BOOLEAN
FatIsCurrentOperationSynchedForDcbTeardown (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb
)
{
PIRP Irp = IrpContext->OriginatingIrp;
PIO_STACK_LOCATION Stack = IoGetCurrentIrpStackLocation( Irp ) ;
PVCB Vcb;
PFCB Fcb;
PCCB Ccb;
PFILE_OBJECT ToCheck[3];
ULONG Index = 0;
PAGED_CODE();
//
// While mounting, we're OK without having to own anything.
//
if (Stack->MajorFunction == IRP_MJ_FILE_SYSTEM_CONTROL &&
Stack->MinorFunction == IRP_MN_MOUNT_VOLUME) {
return TRUE;
}
//
// With the Vcb held, the close path is blocked out.
//
if (ExIsResourceAcquiredSharedLite( &Dcb->Vcb->Resource ) ||
ExIsResourceAcquiredExclusiveLite( &Dcb->Vcb->Resource )) {
return TRUE;
}
//
// Accept this assertion at face value. It comes from GetDirentForFcbOrDcb,
// and is reliable.
//
if (FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_PARENT_BY_CHILD )) {
return TRUE;
}
//
// Determine which fileobjects are around on this operation.
//
if (Stack->MajorFunction == IRP_MJ_SET_INFORMATION &&
Stack->Parameters.SetFile.FileObject) {
ToCheck[Index++] = Stack->Parameters.SetFile.FileObject;
}
if (Stack->FileObject) {
ToCheck[Index++] = Stack->FileObject;
}
ToCheck[Index] = NULL;
//
// If the fileobjects we have are for this dcb or a child of it, we are
// also guaranteed that this dcb isn't going anywhere (even without
// the Vcb).
//
for (Index = 0; ToCheck[Index] != NULL; Index++) {
(VOID) FatDecodeFileObject( ToCheck[Index], &Vcb, &Fcb, &Ccb );
while ( Fcb ) {
if (Fcb == Dcb) {
return TRUE;
}
Fcb = Fcb->ParentDcb;
}
}
return FatDisableParentCheck;
}
#endif // DBG
_Requires_lock_held_(_Global_critical_region_)
VOID
FatOpenDirectoryFile (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb
)
/*++
Routine Description:
This routine opens a new directory file if one is not already open.
Arguments:
Dcb - Pointer to the DCB for the directory
Return Value:
None.
--*/
{
PAGED_CODE();
DebugTrace(+1, Dbg, "FatOpenDirectoryFile\n", 0);
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
//
// If we don't have some hold on this Dcb (there are several ways), there is nothing
// to prevent child files from closing and tearing this branch of the tree down in the
// midst of our slapping this reference onto it.
//
// I really wish we had a proper Fcb synchronization model (like CDFS/UDFS/NTFS).
//
NT_ASSERT( FatIsCurrentOperationSynchedForDcbTeardown( IrpContext, Dcb ));
//
// If we haven't yet set the correct AllocationSize, do so.
//
if (Dcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
FatLookupFileAllocationSize( IrpContext, Dcb );
Dcb->Header.FileSize.LowPart =
Dcb->Header.AllocationSize.LowPart;
}
//
// Setup the Bitmap buffer if it is not big enough already
//
FatCheckFreeDirentBitmap( IrpContext, Dcb );
//
// Check if we need to create a directory file.
//
// We first do a spot check and then synchronize and check again.
//
if (Dcb->Specific.Dcb.DirectoryFile == NULL) {
PFILE_OBJECT DirectoryFileObject = NULL;
FatAcquireDirectoryFileMutex( Dcb->Vcb );
_SEH2_TRY {
if (Dcb->Specific.Dcb.DirectoryFile == NULL) {
PDEVICE_OBJECT RealDevice;
//
// Create the special file object for the directory file, and set
// up its pointers back to the Dcb and the section object pointer.
// Note that setting the DirectoryFile pointer in the Dcb has
// to be the last thing done.
//
// Preallocate a close context since we have no Ccb for this object.
//
RealDevice = Dcb->Vcb->CurrentDevice;
DirectoryFileObject = IoCreateStreamFileObject( NULL, RealDevice );
FatPreallocateCloseContext( Dcb->Vcb);
FatSetFileObject( DirectoryFileObject,
DirectoryFile,
Dcb,
NULL );
//
// Remember this internal open.
//
InterlockedIncrement( (LONG*)&(Dcb->Vcb->InternalOpenCount) );
//
// If this is the root directory, it is also a residual open.
//
if (NodeType( Dcb ) == FAT_NTC_ROOT_DCB) {
InterlockedIncrement( (LONG*)&(Dcb->Vcb->ResidualOpenCount) );
}
DirectoryFileObject->SectionObjectPointer = &Dcb->NonPaged->SectionObjectPointers;
DirectoryFileObject->ReadAccess = TRUE;
DirectoryFileObject->WriteAccess = TRUE;
DirectoryFileObject->DeleteAccess = TRUE;
InterlockedIncrement( (LONG*)&Dcb->Specific.Dcb.DirectoryFileOpenCount );
Dcb->Specific.Dcb.DirectoryFile = DirectoryFileObject;
//
// Indicate we're happy with the fileobject now.
//
DirectoryFileObject = NULL;
}
} _SEH2_FINALLY {
FatReleaseDirectoryFileMutex( Dcb->Vcb );
//
// Rip the object up if we couldn't get the close context.
//
if (DirectoryFileObject) {
ObDereferenceObject( DirectoryFileObject );
}
} _SEH2_END;
}
//
// Finally check if we need to initialize the Cache Map for the
// directory file. The size of the section we are going to map
// the current allocation size for the directory. Note that the
// cache manager will provide syncronization for us.
//
if ( Dcb->Specific.Dcb.DirectoryFile->PrivateCacheMap == NULL ) {
Dcb->Header.ValidDataLength = FatMaxLarge;
Dcb->ValidDataToDisk = MAXULONG;
FatInitializeCacheMap( Dcb->Specific.Dcb.DirectoryFile,
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize,
TRUE,
&FatData.CacheManagerNoOpCallbacks,
Dcb );
}
DebugTrace(-1, Dbg, "FatOpenDirectoryFile -> VOID\n", 0);
return;
}
PFILE_OBJECT
FatOpenEaFile (
IN PIRP_CONTEXT IrpContext,
IN PFCB EaFcb
)
/*++
Routine Description:
This routine opens the Ea file.
Arguments:
EaFcb - Pointer to the Fcb for the Ea file.
Return Value:
Pointer to the new file object.
--*/
{
PFILE_OBJECT EaFileObject = NULL;
PDEVICE_OBJECT RealDevice;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatOpenEaFile\n", 0);
DebugTrace( 0, Dbg, "EaFcb = %p\n", EaFcb);
//
// Create the special file object for the ea file, and set
// up its pointers back to the Fcb and the section object pointer
//
RealDevice = EaFcb->Vcb->CurrentDevice;
EaFileObject = IoCreateStreamFileObject( NULL, RealDevice );
_SEH2_TRY {
FatPreallocateCloseContext( IrpContext->Vcb);
FatSetFileObject( EaFileObject,
EaFile,
EaFcb,
NULL );
//
// Remember this internal, residual open.
//
InterlockedIncrement( (LONG*)&(EaFcb->Vcb->InternalOpenCount) );
InterlockedIncrement( (LONG*)&(EaFcb->Vcb->ResidualOpenCount) );
EaFileObject->SectionObjectPointer = &EaFcb->NonPaged->SectionObjectPointers;
EaFileObject->ReadAccess = TRUE;
EaFileObject->WriteAccess = TRUE;
//
// Finally check if we need to initialize the Cache Map for the
// ea file. The size of the section we are going to map
// the current allocation size for the Fcb.
//
EaFcb->Header.ValidDataLength = FatMaxLarge;
FatInitializeCacheMap( EaFileObject,
(PCC_FILE_SIZES)&EaFcb->Header.AllocationSize,
TRUE,
&FatData.CacheManagerCallbacks,
EaFcb );
CcSetAdditionalCacheAttributes( EaFileObject, TRUE, TRUE );
} _SEH2_FINALLY {
//
// Drop the fileobject if we're raising. Two cases: couldn't get
// the close context, and it is still an UnopenedFileObject, or
// we lost trying to build the cache map - in which case we're
// OK for the close context if we have to.
//
if (_SEH2_AbnormalTermination()) {
ObDereferenceObject( EaFileObject );
}
} _SEH2_END;
DebugTrace(-1, Dbg, "FatOpenEaFile -> %p\n", EaFileObject);
UNREFERENCED_PARAMETER( IrpContext );
return EaFileObject;
}
VOID
FatCloseEaFile (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN BOOLEAN FlushFirst
)
/*++
Routine Description:
This routine shuts down the ea file. Usually this is required when the volume
begins to leave the system: after verify, dismount, deletion, pnp.
Arguments:
Vcb - the volume to close the ea file on
FlushFirst - whether the file should be flushed
Return Value:
None. As a side effect, the EA fileobject in the Vcb is cleared.
Caller must have the Vcb exclusive.
--*/
{
PFILE_OBJECT EaFileObject = Vcb->VirtualEaFile;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatCloseEaFile\n", 0);
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
NT_ASSERT( FatVcbAcquiredExclusive(IrpContext, Vcb) );
if (EaFileObject != NULL) {
EaFileObject = Vcb->VirtualEaFile;
if (FlushFirst) {
CcFlushCache( Vcb->VirtualEaFile->SectionObjectPointer, NULL, 0, NULL );
}
Vcb->VirtualEaFile = NULL;
//
// Empty the Mcb for the Ea file.
//
FatRemoveMcbEntry( Vcb, &Vcb->EaFcb->Mcb, 0, 0xFFFFFFFF );
//
// Uninitialize the cache for this file object and dereference it.
//
FatSyncUninitializeCacheMap( IrpContext, EaFileObject );
ObDereferenceObject( EaFileObject );
}
DebugTrace(-1, Dbg, "FatCloseEaFile -> %p\n", EaFileObject);
}
_Requires_lock_held_(_Global_critical_region_)
VOID
FatSetDirtyBcb (
IN PIRP_CONTEXT IrpContext,
IN PBCB Bcb,
IN PVCB Vcb OPTIONAL,
IN BOOLEAN Reversible
)
/*++
Routine Description:
This routine saves a reference to the bcb in the irp context and
sets the bcb dirty. This will have the affect of keeping the page in
memory until we complete the request
In addition, a DPC is set to fire in 5 seconds (or if one is pending,
pushed back 5 seconds) to mark the volume clean.
Arguments:
Bcb - Supplies the Bcb being set dirty
Vcb - Supplies the volume being marked dirty
Reversible - Supplies TRUE if the specified range of bcb should be repinned
so that the changes can be reversed in a controlled fashion if errors
are encountered.
Return Value:
None.
--*/
{
DebugTrace(+1, Dbg, "FatSetDirtyBcb\n", 0 );
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
DebugTrace( 0, Dbg, "Bcb = %p\n", Bcb );
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb );
//
// Repin the bcb as required
//
if (Reversible) {
FatRepinBcb( IrpContext, Bcb );
}
//
// Set the bcb dirty
//
CcSetDirtyPinnedData( Bcb, NULL );
//
// If volume dirtying isn't disabled for this operation (for
// instance, when we're changing the dirty state), set the
// volume dirty if we were given a Vcb that we want to perform
// clean volume processing on, and return.
//
// As a historical note, we used to key off of the old floppy
// (now deferred flush) bit to disable dirtying behavior. Since
// hotpluggable media can still be yanked while operations are
// in flight, recognize that its really the case that FAT12
// doesn't have the dirty bit.
//
if ( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_DIRTY) &&
ARGUMENT_PRESENT(Vcb) &&
!FatIsFat12(Vcb)) {
KIRQL SavedIrql;
BOOLEAN SetTimer;
LARGE_INTEGER TimeSincePreviousCall;
LARGE_INTEGER CurrentTime;
//
// "Borrow" the irp context spinlock.
//
KeQuerySystemTime( &CurrentTime );
KeAcquireSpinLock( &FatData.GeneralSpinLock, &SavedIrql );
TimeSincePreviousCall.QuadPart =
CurrentTime.QuadPart - Vcb->LastFatMarkVolumeDirtyCall.QuadPart;
//
// If more than one second has elapsed since the prior call
// to here, bump the timer up again and see if we need to
// physically mark the volume dirty.
//
if ( (TimeSincePreviousCall.HighPart != 0) ||
(TimeSincePreviousCall.LowPart > (1000 * 1000 * 10)) ) {
SetTimer = TRUE;
} else {
SetTimer = FALSE;
}
KeReleaseSpinLock( &FatData.GeneralSpinLock, SavedIrql );
if ( SetTimer ) {
LARGE_INTEGER CleanVolumeTimer;
//
// We use a shorter volume clean timer for hot plug volumes.
//
CleanVolumeTimer.QuadPart = FlagOn( Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)
? (LONG)-1500*1000*10
: (LONG)-8*1000*1000*10;
(VOID)KeCancelTimer( &Vcb->CleanVolumeTimer );
(VOID)KeRemoveQueueDpc( &Vcb->CleanVolumeDpc );
//
// We have now synchronized with anybody clearing the dirty
// flag, so we can now see if we really have to actually write
// out the physical bit.
//
if ( !FlagOn(Vcb->VcbState, VCB_STATE_FLAG_VOLUME_DIRTY) ) {
//
// We want to really mark the volume dirty now.
//
if (!FlagOn(Vcb->VcbState, VCB_STATE_FLAG_MOUNTED_DIRTY)) {
FatMarkVolume( IrpContext, Vcb, VolumeDirty );
}
SetFlag( Vcb->VcbState, VCB_STATE_FLAG_VOLUME_DIRTY );
//
// Lock the volume if it is removable.
//
if (FlagOn( Vcb->VcbState, VCB_STATE_FLAG_REMOVABLE_MEDIA)) {
FatToggleMediaEjectDisable( IrpContext, Vcb, TRUE );
}
}
KeAcquireSpinLock( &FatData.GeneralSpinLock, &SavedIrql );
KeQuerySystemTime( &Vcb->LastFatMarkVolumeDirtyCall );
KeReleaseSpinLock( &FatData.GeneralSpinLock, SavedIrql );
KeSetTimer( &Vcb->CleanVolumeTimer,
CleanVolumeTimer,
&Vcb->CleanVolumeDpc );
}
}
DebugTrace(-1, Dbg, "FatSetDirtyBcb -> VOID\n", 0 );
}
VOID
FatRepinBcb (
IN PIRP_CONTEXT IrpContext,
IN PBCB Bcb
)
/*++
Routine Description:
This routine saves a reference to the bcb in the irp context. This will
have the affect of keeping the page in memory until we complete the
request
Arguments:
Bcb - Supplies the Bcb being referenced
Return Value:
None.
--*/
{
PREPINNED_BCBS Repinned;
ULONG i;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatRepinBcb\n", 0 );
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
DebugTrace( 0, Dbg, "Bcb = %p\n", Bcb );
//
// The algorithm is to search the list of repinned records until
// we either find a match for the bcb or we find a null slot.
//
Repinned = &IrpContext->Repinned;
while (TRUE) {
//
// For every entry in the repinned record check if the bcb's
// match or if the entry is null. If the bcb's match then
// we've done because we've already repinned this bcb, if
// the entry is null then we know, because it's densely packed,
// that the bcb is not in the list so add it to the repinned
// record and repin it.
//
for (i = 0; i < REPINNED_BCBS_ARRAY_SIZE; i += 1) {
if (Repinned->Bcb[i] == Bcb) {
DebugTrace(-1, Dbg, "FatRepinBcb -> VOID\n", 0 );
return;
}
if (Repinned->Bcb[i] == NULL) {
Repinned->Bcb[i] = Bcb;
CcRepinBcb( Bcb );
DebugTrace(-1, Dbg, "FatRepinBcb -> VOID\n", 0 );
return;
}
}
//
// We finished checking one repinned record so now locate the next
// repinned record, If there isn't one then allocate and zero out
// a new one.
//
if (Repinned->Next == NULL) {
Repinned->Next = FsRtlAllocatePoolWithTag( PagedPool,
sizeof(REPINNED_BCBS),
TAG_REPINNED_BCB );
RtlZeroMemory( Repinned->Next, sizeof(REPINNED_BCBS) );
}
Repinned = Repinned->Next;
}
}
VOID
FatUnpinRepinnedBcbs (
IN PIRP_CONTEXT IrpContext
)
/*++
Routine Description:
This routine frees all of the repinned bcbs, stored in an IRP context.
Arguments:
Return Value:
None.
--*/
{
IO_STATUS_BLOCK RaiseIosb;
PREPINNED_BCBS Repinned;
BOOLEAN WriteThroughToDisk;
PFILE_OBJECT FileObject = NULL;
BOOLEAN ForceVerify = FALSE;
ULONG i;
PFCB FcbOrDcb = NULL;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatUnpinRepinnedBcbs\n", 0 );
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
//
// The algorithm for this procedure is to scan the entire list of
// repinned records unpinning any repinned bcbs. We start off
// with the first record in the irp context, and while there is a
// record to scan we do the following loop.
//
Repinned = &IrpContext->Repinned;
RaiseIosb.Status = STATUS_SUCCESS;
//
// WinSE bug #307418 "Occasional data corruption when
// standby/resume while copying files to removable FAT
// formatted media".
// Extract main FCB pointer from the irp context - we
// will need it later to detect new file creation operation.
//
if (IrpContext->MajorFunction == IRP_MJ_CREATE &&
IrpContext->OriginatingIrp != NULL) {
PIO_STACK_LOCATION IrpSp;
IrpSp = IoGetCurrentIrpStackLocation( IrpContext->OriginatingIrp );
if (IrpSp != NULL &&
IrpSp->FileObject != NULL &&
IrpSp->FileObject->FsContext != NULL) {
FcbOrDcb = IrpSp->FileObject->FsContext;
}
}
//
// If the request is write through or the media is deferred flush,
// unpin the bcb's write through.
//
WriteThroughToDisk = (BOOLEAN) (!FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_WRITE_THROUGH) &&
IrpContext->Vcb != NULL &&
(FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WRITE_THROUGH) ||
FlagOn(IrpContext->Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)));
while (Repinned != NULL) {
//
// For every non-null entry in the repinned record unpin the
// repinned entry.
//
// If the this is removable media (therefore all requests write-
// through) and the write fails, purge the cache so that we throw
// away the modifications as we will be returning an error to the
// user.
//
for (i = 0; i < REPINNED_BCBS_ARRAY_SIZE; i += 1) {
if (Repinned->Bcb[i] != NULL) {
IO_STATUS_BLOCK Iosb;
if (WriteThroughToDisk &&
FlagOn(IrpContext->Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)) {
FileObject = CcGetFileObjectFromBcb( Repinned->Bcb[i] );
}
CcUnpinRepinnedBcb( Repinned->Bcb[i],
WriteThroughToDisk,
&Iosb );
if (!NT_SUCCESS(Iosb.Status)) {
if (RaiseIosb.Status == STATUS_SUCCESS) {
RaiseIosb = Iosb;
}
//
// If this was a writethrough device, purge the cache,
// except for Irp major codes that either don't handle
// the error paths correctly or are simple victims like
// cleanup.c.
//
if (FileObject &&
(IrpContext->MajorFunction != IRP_MJ_CLEANUP) &&
(IrpContext->MajorFunction != IRP_MJ_FLUSH_BUFFERS) &&
(IrpContext->MajorFunction != IRP_MJ_SET_INFORMATION)
&&
//
// WinSE bug #307418 "Occasional data corruption when
// standby/resume while copying files to removable FAT
// formatted media".
// Buffer unpinning for new file creation operation can
// be interrupted by system syspend. As a result some BCBs
// will be successfully written to the disk while others will
// be kicked back with STATUS_VERIFY_REQUIRED. Since there is
// is still a chance for the failed BCBs to reach the disk
// after the volume verification we'll not purge them.
// Instead FatCommonCreate() will unroll the file creation
// changes for these pages.
//
!(IrpContext->MajorFunction == IRP_MJ_CREATE &&
Iosb.Status == STATUS_VERIFY_REQUIRED &&
FcbOrDcb != NULL &&
NodeType( FcbOrDcb ) == FAT_NTC_FCB)) {
//
// The call to CcPurgeCacheSection() below will
// purge the entire file from memory. It will also
// block until all the file's BCB's are pinned.
//
// We end up in a deadlock situation of there
// are any other pinned BCB's in this IRP context
// so the first thing we do is search the list
// for BCB's pinned in the same file and unpin
// them.
//
// We are probably not going to lose data because
// it's safe to assume that all flushes will
// fail after the first one fails.
//
ULONG j;
ULONG k = i + 1;
PREPINNED_BCBS RepinnedToPurge = Repinned;
while( RepinnedToPurge != NULL ) {
for (j = k; j < REPINNED_BCBS_ARRAY_SIZE; j++) {
if (RepinnedToPurge->Bcb[j] != NULL) {
if (CcGetFileObjectFromBcb( RepinnedToPurge->Bcb[j] ) == FileObject) {
CcUnpinRepinnedBcb( RepinnedToPurge->Bcb[j],
FALSE,
&Iosb );
RepinnedToPurge->Bcb[j] = NULL;
}
}
}
RepinnedToPurge = RepinnedToPurge->Next;
k = 0;
}
CcPurgeCacheSection( FileObject->SectionObjectPointer,
NULL,
0,
FALSE );
//
// Force a verify operation here since who knows
// what state things are in.
//
ForceVerify = TRUE;
}
}
Repinned->Bcb[i] = NULL;
}
}
//
// Now find the next repinned record in the list, and possibly
// delete the one we've just processed.
//
if (Repinned != &IrpContext->Repinned) {
PREPINNED_BCBS Saved;
Saved = Repinned->Next;
ExFreePool( Repinned );
Repinned = Saved;
} else {
Repinned = Repinned->Next;
IrpContext->Repinned.Next = NULL;
}
}
//
// Now if we weren't completely successful in the our unpin
// then raise the iosb we got
//
if (!NT_SUCCESS(RaiseIosb.Status)) {
if (ForceVerify && FileObject) {
SetFlag(FileObject->DeviceObject->Flags, DO_VERIFY_VOLUME);
IoSetHardErrorOrVerifyDevice( IrpContext->OriginatingIrp,
FileObject->DeviceObject );
}
if (!FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_RAISE )) {
if (IrpContext->OriginatingIrp) {
IrpContext->OriginatingIrp->IoStatus = RaiseIosb;
}
FatNormalizeAndRaiseStatus( IrpContext, RaiseIosb.Status );
}
}
DebugTrace(-1, Dbg, "FatUnpinRepinnedBcbs -> VOID\n", 0 );
return;
}
FINISHED
FatZeroData (
IN PIRP_CONTEXT IrpContext,
IN PVCB Vcb,
IN PFILE_OBJECT FileObject,
IN ULONG StartingZero,
IN ULONG ByteCount
)
/*++
**** Temporary function - Remove when CcZeroData is capable of handling
non sector aligned requests.
--*/
{
#ifndef __REACTOS__
LARGE_INTEGER ZeroStart = {0,0};
LARGE_INTEGER BeyondZeroEnd = {0,0};
#else
LARGE_INTEGER ZeroStart = {{0,0}};
LARGE_INTEGER BeyondZeroEnd = {{0,0}};
#endif
ULONG SectorSize;
BOOLEAN Finished;
PAGED_CODE();
SectorSize = (ULONG)Vcb->Bpb.BytesPerSector;
ZeroStart.LowPart = (StartingZero + (SectorSize - 1)) & ~(SectorSize - 1);
//
// Detect overflow if we were asked to zero in the last sector of the file,
// which must be "zeroed" already (or we're in trouble).
//
if (StartingZero != 0 && ZeroStart.LowPart == 0) {
return TRUE;
}
//
// Note that BeyondZeroEnd can take the value 4gb.
//
BeyondZeroEnd.QuadPart = ((ULONGLONG) StartingZero + ByteCount + (SectorSize - 1))
& (~((LONGLONG) SectorSize - 1));
//
// If we were called to just zero part of a sector we are in trouble.
//
if ( ZeroStart.QuadPart == BeyondZeroEnd.QuadPart ) {
return TRUE;
}
Finished = CcZeroData( FileObject,
&ZeroStart,
&BeyondZeroEnd,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
return Finished;
}
NTSTATUS
FatCompleteMdl (
IN PIRP_CONTEXT IrpContext,
IN PIRP Irp
)
/*++
Routine Description:
This routine performs the function of completing Mdl read and write
requests. It should be called only from FatFsdRead and FatFsdWrite.
Arguments:
Irp - Supplies the originating Irp.
Return Value:
NTSTATUS - Will always be STATUS_PENDING or STATUS_SUCCESS.
--*/
{
PFILE_OBJECT FileObject;
PIO_STACK_LOCATION IrpSp;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatCompleteMdl\n", 0 );
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
DebugTrace( 0, Dbg, "Irp = %p\n", Irp );
//
// Do completion processing.
//
FileObject = IoGetCurrentIrpStackLocation( Irp )->FileObject;
switch( IrpContext->MajorFunction ) {
case IRP_MJ_READ:
CcMdlReadComplete( FileObject, Irp->MdlAddress );
break;
case IRP_MJ_WRITE:
IrpSp = IoGetCurrentIrpStackLocation( Irp );
NT_ASSERT( FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT ));
CcMdlWriteComplete( FileObject, &IrpSp->Parameters.Write.ByteOffset, Irp->MdlAddress );
Irp->IoStatus.Status = STATUS_SUCCESS;
break;
default:
DebugTrace( DEBUG_TRACE_ERROR, 0, "Illegal Mdl Complete.\n", 0);
#ifdef _MSC_VER
#pragma prefast( suppress: 28159, "we're very broken if we get here" )
#endif
FatBugCheck( IrpContext->MajorFunction, 0, 0 );
}
//
// Mdl is now deallocated.
//
Irp->MdlAddress = NULL;
//
// Complete the request and exit right away.
//
FatCompleteRequest( IrpContext, Irp, STATUS_SUCCESS );
DebugTrace(-1, Dbg, "FatCompleteMdl -> STATUS_SUCCESS\n", 0 );
return STATUS_SUCCESS;
}
VOID
FatSyncUninitializeCacheMap (
IN PIRP_CONTEXT IrpContext,
IN PFILE_OBJECT FileObject
)
/*++
Routine Description:
The routine performs a CcUnitializeCacheMap to LargeZero synchronously. That
is it waits on the Cc event. This call is useful when we want to be certain
when a close will actually some in.
Return Value:
None.
--*/
{
CACHE_UNINITIALIZE_EVENT UninitializeCompleteEvent;
NTSTATUS WaitStatus;
UNREFERENCED_PARAMETER( IrpContext );
PAGED_CODE();
KeInitializeEvent( &UninitializeCompleteEvent.Event,
SynchronizationEvent,
FALSE);
CcUninitializeCacheMap( FileObject,
&FatLargeZero,
&UninitializeCompleteEvent );
//
// Now wait for the cache manager to finish purging the file.
// This will garentee that Mm gets the purge before we
// delete the Vcb.
//
#ifdef _MSC_VER
#pragma prefast( suppress: 28931, "we use WaitStatus in the debug assert, in fre builds prefast complains it's unused" )
#endif
WaitStatus = KeWaitForSingleObject( &UninitializeCompleteEvent.Event,
Executive,
KernelMode,
FALSE,
NULL);
NT_ASSERT(WaitStatus == STATUS_SUCCESS);
}
VOID
FatPinMappedData (
IN PIRP_CONTEXT IrpContext,
IN PDCB Dcb,
IN VBO StartingVbo,
IN ULONG ByteCount,
OUT PBCB *Bcb
)
/*++
Routine Description:
This routine pins data that was previously mapped before setting it dirty.
Arguments:
Dcb - Pointer to the DCB for the directory
StartingVbo - The virtual offset of the first desired byte
ByteCount - Number of bytes desired
Bcb - Returns a pointer to the BCB which is valid until unpinned
--*/
{
LARGE_INTEGER Vbo;
PAGED_CODE();
DebugTrace(+1, Dbg, "FatPinMappedData\n", 0);
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
//
// Call the Cache manager to perform the operation.
//
Vbo.QuadPart = StartingVbo;
if (!CcPinMappedData( Dcb->Specific.Dcb.DirectoryFile,
&Vbo,
ByteCount,
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
Bcb )) {
//
// Could not pin the data without waiting (cache miss).
//
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
}
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID, *BCB = %p\n", *Bcb);
return;
}
#if (NTDDI_VERSION >= NTDDI_WIN8)
NTSTATUS
FatPrefetchPages (
IN PIRP_CONTEXT IrpContext,
IN PFILE_OBJECT FileObject,
IN ULONG StartingPage,
IN ULONG PageCount
)
{
IO_PRIORITY_INFO PriorityInformation = {0};
MM_PREFETCH_FLAGS PrefetchFlags;
ULONG PageNo;
NTSTATUS Status;
PREAD_LIST ReadList = NULL;
UNREFERENCED_PARAMETER( IrpContext );
PAGED_CODE();
//
// Succeed zero page prefetch requests.
//
if (PageCount == 0) {
return STATUS_SUCCESS;
}
//
// Mm's prefetch API's "only" support fetching a ULONG worth of pages.
// Make sure we don't overflow.
//
ASSERT( PageCount < (PFN_NUMBER)MAXULONG );
IoInitializePriorityInfo( &PriorityInformation );
Status = IoRetrievePriorityInfo( IrpContext->OriginatingIrp,
FileObject,
IrpContext->OriginatingIrp->Tail.Overlay.Thread,
&PriorityInformation );
if (!NT_SUCCESS( Status)) {
goto Cleanup;
}
ReadList = ExAllocatePoolWithTag( PagedPool,
FIELD_OFFSET( READ_LIST, List ) + PageCount * sizeof( FILE_SEGMENT_ELEMENT ),
' taF' );
if (ReadList == NULL) {
Status = STATUS_INSUFFICIENT_RESOURCES;
goto Cleanup;
}
//
// Call Mm to prefetch data.
//
ReadList->FileObject = FileObject;
ReadList->IsImage = FALSE;
ReadList->NumberOfEntries = PageCount;
PrefetchFlags.AllFlags = 0;
PrefetchFlags.Flags.Priority = PriorityInformation.PagePriority;
PrefetchFlags.Flags.RepurposePriority = SYSTEM_PAGE_PRIORITY_LEVELS - 1;
PrefetchFlags.Flags.PriorityProtection = 1;
ReadList->List[0].Alignment = StartingPage * PAGE_SIZE;
ReadList->List[0].Alignment |= PrefetchFlags.AllFlags;
for (PageNo = 1; PageNo < PageCount; PageNo++) {
ReadList->List[PageNo].Alignment = ReadList->List[PageNo-1].Alignment + PAGE_SIZE;
}
Status = MmPrefetchPages( 1, &ReadList );
Cleanup:
if (ReadList != NULL) {
ExFreePoolWithTag( ReadList, ' taF' );
}
return Status;
}
#endif