mirror of
https://github.com/reactos/reactos.git
synced 2024-12-29 10:35:28 +00:00
2017 lines
52 KiB
C
2017 lines
52 KiB
C
/*++
|
|
|
|
Copyright (c) 1990-2000 Microsoft Corporation
|
|
|
|
Module Name:
|
|
|
|
cache.c
|
|
|
|
Abstract:
|
|
|
|
This module implements the cache management routines for the Fat
|
|
FSD and FSP, by calling the Common Cache Manager.
|
|
|
|
|
|
--*/
|
|
|
|
#include "fatprocs.h"
|
|
|
|
//
|
|
// The Bug check file id for this module
|
|
//
|
|
|
|
#define BugCheckFileId (FAT_BUG_CHECK_CACHESUP)
|
|
|
|
//
|
|
// Local debug trace level
|
|
//
|
|
|
|
#define Dbg (DEBUG_TRACE_CACHESUP)
|
|
|
|
#if DBG
|
|
|
|
BOOLEAN
|
|
FatIsCurrentOperationSynchedForDcbTeardown (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb
|
|
);
|
|
|
|
#endif
|
|
|
|
#ifdef ALLOC_PRAGMA
|
|
#pragma alloc_text(PAGE, FatCloseEaFile)
|
|
#pragma alloc_text(PAGE, FatCompleteMdl)
|
|
#pragma alloc_text(PAGE, FatOpenDirectoryFile)
|
|
#pragma alloc_text(PAGE, FatOpenEaFile)
|
|
#pragma alloc_text(PAGE, FatPinMappedData)
|
|
#pragma alloc_text(PAGE, FatPrepareWriteDirectoryFile)
|
|
#pragma alloc_text(PAGE, FatPrepareWriteVolumeFile)
|
|
#pragma alloc_text(PAGE, FatReadDirectoryFile)
|
|
#pragma alloc_text(PAGE, FatReadVolumeFile)
|
|
#pragma alloc_text(PAGE, FatRepinBcb)
|
|
#pragma alloc_text(PAGE, FatSyncUninitializeCacheMap)
|
|
#pragma alloc_text(PAGE, FatUnpinRepinnedBcbs)
|
|
#pragma alloc_text(PAGE, FatZeroData)
|
|
#pragma alloc_text(PAGE, FatPrefetchPages)
|
|
#if DBG
|
|
#pragma alloc_text(PAGE, FatIsCurrentOperationSynchedForDcbTeardown)
|
|
#endif
|
|
#endif
|
|
|
|
VOID
|
|
FatInitializeCacheMap (
|
|
_In_ PFILE_OBJECT FileObject,
|
|
_In_ PCC_FILE_SIZES FileSizes,
|
|
_In_ BOOLEAN PinAccess,
|
|
_In_ PCACHE_MANAGER_CALLBACKS Callbacks,
|
|
_In_ PVOID LazyWriteContext
|
|
)
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
Wrapper over CcInitializeCacheMap and CcSetAdditionalCacheAttributesEx to initialize
|
|
caching and enable IO accounting on a file.
|
|
|
|
--*/
|
|
|
|
{
|
|
//
|
|
// Initialize caching
|
|
//
|
|
|
|
CcInitializeCacheMap( FileObject,
|
|
FileSizes,
|
|
PinAccess,
|
|
Callbacks,
|
|
LazyWriteContext );
|
|
|
|
#if (NTDDI_VERSION >= NTDDI_WIN8)
|
|
//
|
|
// Enable Disk IO Accounting for this file
|
|
//
|
|
|
|
if (FatDiskAccountingEnabled) {
|
|
|
|
CcSetAdditionalCacheAttributesEx( FileObject, CC_ENABLE_DISK_IO_ACCOUNTING );
|
|
}
|
|
#endif
|
|
}
|
|
|
|
VOID
|
|
FatReadVolumeFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PVCB Vcb,
|
|
IN VBO StartingVbo,
|
|
IN ULONG ByteCount,
|
|
OUT PBCB *Bcb,
|
|
OUT PVOID *Buffer
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine is called when the specified range of sectors is to be
|
|
read into the cache. In fat, the volume file only contains the boot
|
|
sector, reserved sectors, and the "fat(s)." Thus the volume file is
|
|
of fixed size and only extends up to (but not not including) the root
|
|
directory entry, and will never move or change size.
|
|
|
|
The fat volume file is also peculiar in that, since it starts at the
|
|
logical beginning of the disk, Vbo == Lbo.
|
|
|
|
Arguments:
|
|
|
|
Vcb - Pointer to the VCB for the volume
|
|
|
|
StartingVbo - The virtual offset of the first desired byte
|
|
|
|
ByteCount - Number of bytes desired
|
|
|
|
Bcb - Returns a pointer to the BCB which is valid until unpinned
|
|
|
|
Buffer - Returns a pointer to the sectors, which is valid until unpinned
|
|
|
|
--*/
|
|
|
|
{
|
|
LARGE_INTEGER Vbo;
|
|
|
|
PAGED_CODE();
|
|
|
|
//
|
|
// Check to see that all references are within the Bios Parameter Block
|
|
// or the fat(s). A special case is made when StartingVbo == 0 at
|
|
// mounting time since we do not know how big the fat is.
|
|
//
|
|
|
|
NT_ASSERT( ((StartingVbo == 0) || ((StartingVbo + ByteCount) <= (ULONG)
|
|
(FatRootDirectoryLbo( &Vcb->Bpb ) + PAGE_SIZE))));
|
|
|
|
DebugTrace(+1, Dbg, "FatReadVolumeFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
|
|
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
|
|
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
|
|
|
|
//
|
|
// Call the Cache manager to attempt the transfer.
|
|
//
|
|
|
|
Vbo.QuadPart = StartingVbo;
|
|
|
|
if (!CcMapData( Vcb->VirtualVolumeFile,
|
|
&Vbo,
|
|
ByteCount,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
Bcb,
|
|
Buffer )) {
|
|
|
|
NT_ASSERT( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
|
|
|
|
//
|
|
// Could not read the data without waiting (cache miss).
|
|
//
|
|
|
|
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
|
|
}
|
|
|
|
DbgDoit( IrpContext->PinCount += 1 )
|
|
|
|
DebugTrace(-1, Dbg, "FatReadVolumeFile -> VOID, *BCB = %p\n", *Bcb);
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
_Requires_lock_held_(_Global_critical_region_)
|
|
VOID
|
|
FatPrepareWriteVolumeFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PVCB Vcb,
|
|
IN VBO StartingVbo,
|
|
IN ULONG ByteCount,
|
|
OUT PBCB *Bcb,
|
|
OUT PVOID *Buffer,
|
|
IN BOOLEAN Reversible,
|
|
IN BOOLEAN Zero
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine first looks to see if the specified range of sectors,
|
|
is already in the cache. If so, it increments the BCB PinCount,
|
|
sets the BCB dirty, and returns with the location of the sectors.
|
|
|
|
If the sectors are not in the cache and Wait is TRUE, it finds a
|
|
free BCB (potentially causing a flush), and clears out the entire
|
|
buffer. Once this is done, it increments the BCB PinCount, sets the
|
|
BCB dirty, and returns with the location of the sectors.
|
|
|
|
If the sectors are not in the cache and Wait is FALSE, this routine
|
|
raises STATUS_CANT_WAIT.
|
|
|
|
Arguments:
|
|
|
|
Vcb - Pointer to the VCB for the volume
|
|
|
|
StartingVbo - The virtual offset of the first byte to be written
|
|
|
|
ByteCount - Number of bytes to be written
|
|
|
|
Bcb - Returns a pointer to the BCB which is valid until unpinned
|
|
|
|
Buffer - Returns a pointer to the sectors, which is valid until unpinned
|
|
|
|
Reversible - Supplies TRUE if the specified range of modification should
|
|
be repinned so that the operation can be reversed in a controlled
|
|
fashion if errors are encountered.
|
|
|
|
Zero - Supplies TRUE if the specified range of bytes should be zeroed
|
|
|
|
--*/
|
|
|
|
{
|
|
LARGE_INTEGER Vbo;
|
|
|
|
PAGED_CODE();
|
|
|
|
//
|
|
// Check to see that all references are within the Bios Parameter Block
|
|
// or the fat(s).
|
|
//
|
|
|
|
NT_ASSERT( ((StartingVbo + ByteCount) <= (ULONG)
|
|
(FatRootDirectoryLbo( &Vcb->Bpb ))));
|
|
|
|
DebugTrace(+1, Dbg, "FatPrepareWriteVolumeFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
|
|
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", (ULONG)StartingVbo);
|
|
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
|
|
DebugTrace( 0, Dbg, "Zero = %08lx\n", Zero);
|
|
|
|
//
|
|
// Call the Cache manager to attempt the transfer.
|
|
//
|
|
|
|
Vbo.QuadPart = StartingVbo;
|
|
|
|
if (!CcPinRead( Vcb->VirtualVolumeFile,
|
|
&Vbo,
|
|
ByteCount,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
Bcb,
|
|
Buffer )) {
|
|
|
|
NT_ASSERT( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
|
|
|
|
//
|
|
// Could not read the data without waiting (cache miss).
|
|
//
|
|
|
|
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
|
|
}
|
|
|
|
//
|
|
// This keeps the data pinned until we complete the request
|
|
// and writes the dirty bit through to the disk.
|
|
//
|
|
|
|
DbgDoit( IrpContext->PinCount += 1 )
|
|
|
|
_SEH2_TRY {
|
|
|
|
if (Zero) {
|
|
|
|
RtlZeroMemory( *Buffer, ByteCount );
|
|
}
|
|
|
|
FatSetDirtyBcb( IrpContext, *Bcb, Vcb, Reversible );
|
|
|
|
} _SEH2_FINALLY {
|
|
|
|
if (_SEH2_AbnormalTermination()) {
|
|
|
|
FatUnpinBcb(IrpContext, *Bcb);
|
|
}
|
|
} _SEH2_END;
|
|
|
|
DebugTrace(-1, Dbg, "FatPrepareWriteVolumeFile -> VOID, *Bcb = %p\n", *Bcb);
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
_Requires_lock_held_(_Global_critical_region_)
|
|
VOID
|
|
FatReadDirectoryFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb,
|
|
IN VBO StartingVbo,
|
|
IN ULONG ByteCount,
|
|
IN BOOLEAN Pin,
|
|
OUT PBCB *Bcb,
|
|
OUT PVOID *Buffer,
|
|
OUT PNTSTATUS Status
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine is called when the specified range of sectors is to be
|
|
read into the cache. If the desired range falls beyond the current
|
|
cache mapping, the fat will be searched, and if the desired range can
|
|
be satisfied, the cache mapping will be extended and the MCB updated
|
|
accordingly.
|
|
|
|
Arguments:
|
|
|
|
Dcb - Pointer to the DCB for the directory
|
|
|
|
StartingVbo - The virtual offset of the first desired byte
|
|
|
|
ByteCount - Number of bytes desired
|
|
|
|
Pin - Tells us if we should pin instead of just mapping.
|
|
|
|
Bcb - Returns a pointer to the BCB which is valid until unpinned
|
|
|
|
Buffer - Returns a pointer to the sectors, which is valid until unpinned
|
|
|
|
Status - Returns the status of the operation.
|
|
|
|
--*/
|
|
|
|
{
|
|
LARGE_INTEGER Vbo;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatReadDirectoryFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
|
|
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
|
|
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
|
|
|
|
//
|
|
// Check for the zero case
|
|
//
|
|
|
|
if (ByteCount == 0) {
|
|
|
|
DebugTrace(0, Dbg, "Nothing to read\n", 0);
|
|
|
|
*Bcb = NULL;
|
|
*Buffer = NULL;
|
|
*Status = STATUS_SUCCESS;
|
|
|
|
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID\n", 0);
|
|
return;
|
|
}
|
|
|
|
//
|
|
// If we need to create a directory file and initialize the
|
|
// cachemap, do so.
|
|
//
|
|
|
|
FatOpenDirectoryFile( IrpContext, Dcb );
|
|
|
|
//
|
|
// Now if the transfer is beyond the allocation size return EOF.
|
|
//
|
|
|
|
if (StartingVbo >= Dcb->Header.AllocationSize.LowPart) {
|
|
|
|
DebugTrace(0, Dbg, "End of file read for directory\n", 0);
|
|
|
|
*Bcb = NULL;
|
|
*Buffer = NULL;
|
|
*Status = STATUS_END_OF_FILE;
|
|
|
|
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID\n", 0);
|
|
return;
|
|
}
|
|
|
|
//
|
|
// If the caller is trying to read past the EOF, truncate the
|
|
// read.
|
|
//
|
|
|
|
ByteCount = (Dcb->Header.AllocationSize.LowPart - StartingVbo < ByteCount) ?
|
|
Dcb->Header.AllocationSize.LowPart - StartingVbo : ByteCount;
|
|
|
|
NT_ASSERT( ByteCount != 0 );
|
|
|
|
//
|
|
// Call the Cache manager to attempt the transfer.
|
|
//
|
|
|
|
Vbo.QuadPart = StartingVbo;
|
|
|
|
if (Pin ?
|
|
|
|
!CcPinRead( Dcb->Specific.Dcb.DirectoryFile,
|
|
&Vbo,
|
|
ByteCount,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
Bcb,
|
|
Buffer )
|
|
:
|
|
|
|
!CcMapData( Dcb->Specific.Dcb.DirectoryFile,
|
|
&Vbo,
|
|
ByteCount,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
Bcb,
|
|
Buffer ) ) {
|
|
|
|
//
|
|
// Could not read the data without waiting (cache miss).
|
|
//
|
|
|
|
*Bcb = NULL;
|
|
*Buffer = NULL;
|
|
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
|
|
}
|
|
|
|
DbgDoit( IrpContext->PinCount += 1 )
|
|
|
|
*Status = STATUS_SUCCESS;
|
|
|
|
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID, *BCB = %p\n", *Bcb);
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
_Requires_lock_held_(_Global_critical_region_)
|
|
VOID
|
|
FatPrepareWriteDirectoryFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb,
|
|
IN VBO StartingVbo,
|
|
IN ULONG ByteCount,
|
|
OUT PBCB *Bcb,
|
|
OUT PVOID *Buffer,
|
|
IN BOOLEAN Zero,
|
|
IN BOOLEAN Reversible,
|
|
OUT PNTSTATUS Status
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine first looks to see if the specified range of sectors
|
|
is already in the cache. If so, it increments the BCB PinCount,
|
|
sets the BCB dirty, and returns TRUE with the location of the sectors.
|
|
|
|
The IrpContext->Flags .. Wait == TRUE/FALSE actions of this routine are identical to
|
|
FatPrepareWriteVolumeFile() above.
|
|
|
|
Arguments:
|
|
|
|
Dcb - Pointer to the DCB for the directory
|
|
|
|
StartingVbo - The virtual offset of the first byte to be written
|
|
|
|
ByteCount - Number of bytes to be written
|
|
|
|
Bcb - Returns a pointer to the BCB which is valid until unpinned
|
|
|
|
Buffer - Returns a pointer to the sectors, which is valid until unpinned
|
|
|
|
Zero - Supplies TRUE if the specified range of bytes should be zeroed
|
|
|
|
Reversible - Supplies TRUE if the specified range of modification should
|
|
be repinned so that the operation can be reversed in a controlled
|
|
fashion if errors are encountered.
|
|
|
|
Status - Returns the status of the operation.
|
|
|
|
--*/
|
|
|
|
{
|
|
LARGE_INTEGER Vbo;
|
|
ULONG InitialAllocation = 0;
|
|
BOOLEAN UnwindWeAllocatedDiskSpace = FALSE;
|
|
PBCB LocalBcb = NULL;
|
|
PVOID LocalBuffer = NULL;
|
|
ULONG InitialRequest = ByteCount;
|
|
ULONG MappingGranularity = PAGE_SIZE;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatPrepareWriteDirectoryFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
|
|
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", (ULONG)StartingVbo);
|
|
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
|
|
DebugTrace( 0, Dbg, "Zero = %08lx\n", Zero);
|
|
|
|
*Bcb = NULL;
|
|
*Buffer = NULL;
|
|
|
|
//
|
|
// If we need to create a directory file and initialize the
|
|
// cachemap, do so.
|
|
//
|
|
|
|
FatOpenDirectoryFile( IrpContext, Dcb );
|
|
|
|
//
|
|
// If the transfer is beyond the allocation size we need to
|
|
// extend the directory's allocation. The call to
|
|
// AddFileAllocation will raise a condition if
|
|
// it runs out of disk space. Note that the root directory
|
|
// cannot be extended.
|
|
//
|
|
|
|
Vbo.QuadPart = StartingVbo;
|
|
|
|
_SEH2_TRY {
|
|
|
|
if (StartingVbo + ByteCount > Dcb->Header.AllocationSize.LowPart) {
|
|
|
|
if (NodeType(Dcb) == FAT_NTC_ROOT_DCB &&
|
|
!FatIsFat32(Dcb->Vcb)) {
|
|
|
|
FatRaiseStatus( IrpContext, STATUS_DISK_FULL );
|
|
}
|
|
|
|
DebugTrace(0, Dbg, "Try extending normal directory\n", 0);
|
|
|
|
InitialAllocation = Dcb->Header.AllocationSize.LowPart;
|
|
|
|
FatAddFileAllocation( IrpContext,
|
|
Dcb,
|
|
Dcb->Specific.Dcb.DirectoryFile,
|
|
StartingVbo + ByteCount );
|
|
|
|
UnwindWeAllocatedDiskSpace = TRUE;
|
|
|
|
//
|
|
// Inform the cache manager of the new allocation
|
|
//
|
|
|
|
Dcb->Header.FileSize.LowPart =
|
|
Dcb->Header.AllocationSize.LowPart;
|
|
|
|
CcSetFileSizes( Dcb->Specific.Dcb.DirectoryFile,
|
|
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize );
|
|
|
|
//
|
|
// Set up the Bitmap buffer if it is not big enough already
|
|
//
|
|
|
|
FatCheckFreeDirentBitmap( IrpContext, Dcb );
|
|
|
|
//
|
|
// The newly allocated clusters should be zeroed starting at
|
|
// the previous allocation size
|
|
//
|
|
|
|
Zero = TRUE;
|
|
Vbo.QuadPart = InitialAllocation;
|
|
ByteCount = Dcb->Header.AllocationSize.LowPart - InitialAllocation;
|
|
}
|
|
|
|
while (ByteCount > 0) {
|
|
|
|
ULONG BytesToPin;
|
|
|
|
LocalBcb = NULL;
|
|
|
|
//
|
|
// We must pin in terms of pages below the boundary of the initial request.
|
|
// Once we pass the end of the request, we are free to expand the pin size to
|
|
// VACB_MAPPING_GRANULARITY. This will prevent Cc from returning OBCBs
|
|
// and hence will prevent bugchecks when we then attempt to repin one, yet
|
|
// allow us to be more efficient by pinning in 256KB chunks instead of 4KB pages.
|
|
//
|
|
|
|
if (Vbo.QuadPart > StartingVbo + InitialRequest) {
|
|
|
|
MappingGranularity = VACB_MAPPING_GRANULARITY;
|
|
}
|
|
|
|
//
|
|
// If the first and final byte are both described by the same page, pin
|
|
// the entire range. Note we pin in pages to prevent cache manager from
|
|
// returning OBCBs, which would result in a bugcheck on CcRepinBcb.
|
|
//
|
|
|
|
if ((Vbo.QuadPart / MappingGranularity) ==
|
|
((Vbo.QuadPart + ByteCount - 1) / MappingGranularity)) {
|
|
|
|
BytesToPin = ByteCount;
|
|
|
|
} else {
|
|
|
|
BytesToPin = MappingGranularity -
|
|
((ULONG)Vbo.QuadPart & (MappingGranularity - 1));
|
|
}
|
|
|
|
if (!CcPinRead( Dcb->Specific.Dcb.DirectoryFile,
|
|
&Vbo,
|
|
BytesToPin,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
&LocalBcb,
|
|
&LocalBuffer )) {
|
|
|
|
//
|
|
// Could not read the data without waiting (cache miss).
|
|
//
|
|
|
|
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
|
|
}
|
|
|
|
//
|
|
// Update our caller with the beginning of their request.
|
|
//
|
|
|
|
if (*Buffer == NULL) {
|
|
|
|
*Buffer = LocalBuffer;
|
|
*Bcb = LocalBcb;
|
|
}
|
|
|
|
DbgDoit( IrpContext->PinCount += 1 )
|
|
|
|
if (Zero) {
|
|
|
|
//
|
|
// We set this guy dirty right now so that we can raise CANT_WAIT when
|
|
// it needs to be done. It'd be beautiful if we could noop the read IO
|
|
// since we know we don't care about it.
|
|
//
|
|
|
|
RtlZeroMemory( LocalBuffer, BytesToPin );
|
|
CcSetDirtyPinnedData( LocalBcb, NULL );
|
|
}
|
|
|
|
ByteCount -= BytesToPin;
|
|
Vbo.QuadPart += BytesToPin;
|
|
|
|
if (*Bcb != LocalBcb) {
|
|
|
|
FatRepinBcb( IrpContext, LocalBcb );
|
|
FatUnpinBcb( IrpContext, LocalBcb );
|
|
}
|
|
}
|
|
|
|
//
|
|
// This lets us get the data pinned until we complete the request
|
|
// and writes the dirty bit through to the disk.
|
|
//
|
|
|
|
FatSetDirtyBcb( IrpContext, *Bcb, Dcb->Vcb, Reversible );
|
|
|
|
*Status = STATUS_SUCCESS;
|
|
|
|
} _SEH2_FINALLY {
|
|
|
|
DebugUnwind( FatPrepareWriteDirectoryFile );
|
|
|
|
if (_SEH2_AbnormalTermination()) {
|
|
|
|
//
|
|
// Make sure we unpin the buffers.
|
|
//
|
|
|
|
if (*Bcb != LocalBcb) {
|
|
|
|
FatUnpinBcb( IrpContext, LocalBcb );
|
|
}
|
|
|
|
FatUnpinBcb(IrpContext, *Bcb);
|
|
|
|
//
|
|
// These steps are carefully arranged - FatTruncateFileAllocation can raise.
|
|
// Make sure we unpin the buffer. If FTFA raises, the effect should be benign.
|
|
//
|
|
|
|
if (UnwindWeAllocatedDiskSpace == TRUE) {
|
|
|
|
//
|
|
// Inform the cache manager of the change.
|
|
//
|
|
|
|
FatTruncateFileAllocation( IrpContext, Dcb, InitialAllocation );
|
|
|
|
Dcb->Header.FileSize.LowPart =
|
|
Dcb->Header.AllocationSize.LowPart;
|
|
|
|
CcSetFileSizes( Dcb->Specific.Dcb.DirectoryFile,
|
|
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize );
|
|
}
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatPrepareWriteDirectoryFile -> (VOID), *Bcb = %p\n", *Bcb);
|
|
} _SEH2_END;
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
#if DBG
|
|
BOOLEAN FatDisableParentCheck = 0;
|
|
|
|
BOOLEAN
|
|
FatIsCurrentOperationSynchedForDcbTeardown (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb
|
|
)
|
|
{
|
|
PIRP Irp = IrpContext->OriginatingIrp;
|
|
PIO_STACK_LOCATION Stack = IoGetCurrentIrpStackLocation( Irp ) ;
|
|
PVCB Vcb;
|
|
PFCB Fcb;
|
|
PCCB Ccb;
|
|
|
|
PFILE_OBJECT ToCheck[3];
|
|
ULONG Index = 0;
|
|
|
|
PAGED_CODE();
|
|
|
|
//
|
|
// While mounting, we're OK without having to own anything.
|
|
//
|
|
|
|
if (Stack->MajorFunction == IRP_MJ_FILE_SYSTEM_CONTROL &&
|
|
Stack->MinorFunction == IRP_MN_MOUNT_VOLUME) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
//
|
|
// With the Vcb held, the close path is blocked out.
|
|
//
|
|
|
|
if (ExIsResourceAcquiredSharedLite( &Dcb->Vcb->Resource ) ||
|
|
ExIsResourceAcquiredExclusiveLite( &Dcb->Vcb->Resource )) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
//
|
|
// Accept this assertion at face value. It comes from GetDirentForFcbOrDcb,
|
|
// and is reliable.
|
|
//
|
|
|
|
if (FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_PARENT_BY_CHILD )) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
//
|
|
// Determine which fileobjects are around on this operation.
|
|
//
|
|
|
|
if (Stack->MajorFunction == IRP_MJ_SET_INFORMATION &&
|
|
Stack->Parameters.SetFile.FileObject) {
|
|
|
|
ToCheck[Index++] = Stack->Parameters.SetFile.FileObject;
|
|
}
|
|
|
|
if (Stack->FileObject) {
|
|
|
|
ToCheck[Index++] = Stack->FileObject;
|
|
}
|
|
|
|
ToCheck[Index] = NULL;
|
|
|
|
//
|
|
// If the fileobjects we have are for this dcb or a child of it, we are
|
|
// also guaranteed that this dcb isn't going anywhere (even without
|
|
// the Vcb).
|
|
//
|
|
|
|
for (Index = 0; ToCheck[Index] != NULL; Index++) {
|
|
|
|
(VOID) FatDecodeFileObject( ToCheck[Index], &Vcb, &Fcb, &Ccb );
|
|
|
|
while ( Fcb ) {
|
|
|
|
if (Fcb == Dcb) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
Fcb = Fcb->ParentDcb;
|
|
}
|
|
}
|
|
|
|
return FatDisableParentCheck;
|
|
}
|
|
#endif // DBG
|
|
|
|
_Requires_lock_held_(_Global_critical_region_)
|
|
VOID
|
|
FatOpenDirectoryFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine opens a new directory file if one is not already open.
|
|
|
|
Arguments:
|
|
|
|
Dcb - Pointer to the DCB for the directory
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
|
|
{
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatOpenDirectoryFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
|
|
|
|
//
|
|
// If we don't have some hold on this Dcb (there are several ways), there is nothing
|
|
// to prevent child files from closing and tearing this branch of the tree down in the
|
|
// midst of our slapping this reference onto it.
|
|
//
|
|
// I really wish we had a proper Fcb synchronization model (like CDFS/UDFS/NTFS).
|
|
//
|
|
|
|
NT_ASSERT( FatIsCurrentOperationSynchedForDcbTeardown( IrpContext, Dcb ));
|
|
|
|
//
|
|
// If we haven't yet set the correct AllocationSize, do so.
|
|
//
|
|
|
|
if (Dcb->Header.AllocationSize.QuadPart == FCB_LOOKUP_ALLOCATIONSIZE_HINT) {
|
|
|
|
FatLookupFileAllocationSize( IrpContext, Dcb );
|
|
|
|
Dcb->Header.FileSize.LowPart =
|
|
Dcb->Header.AllocationSize.LowPart;
|
|
}
|
|
|
|
//
|
|
// Setup the Bitmap buffer if it is not big enough already
|
|
//
|
|
|
|
FatCheckFreeDirentBitmap( IrpContext, Dcb );
|
|
|
|
//
|
|
// Check if we need to create a directory file.
|
|
//
|
|
// We first do a spot check and then synchronize and check again.
|
|
//
|
|
|
|
if (Dcb->Specific.Dcb.DirectoryFile == NULL) {
|
|
|
|
PFILE_OBJECT DirectoryFileObject = NULL;
|
|
|
|
FatAcquireDirectoryFileMutex( Dcb->Vcb );
|
|
|
|
_SEH2_TRY {
|
|
|
|
if (Dcb->Specific.Dcb.DirectoryFile == NULL) {
|
|
|
|
PDEVICE_OBJECT RealDevice;
|
|
|
|
//
|
|
// Create the special file object for the directory file, and set
|
|
// up its pointers back to the Dcb and the section object pointer.
|
|
// Note that setting the DirectoryFile pointer in the Dcb has
|
|
// to be the last thing done.
|
|
//
|
|
// Preallocate a close context since we have no Ccb for this object.
|
|
//
|
|
|
|
RealDevice = Dcb->Vcb->CurrentDevice;
|
|
|
|
DirectoryFileObject = IoCreateStreamFileObject( NULL, RealDevice );
|
|
FatPreallocateCloseContext( Dcb->Vcb);
|
|
|
|
FatSetFileObject( DirectoryFileObject,
|
|
DirectoryFile,
|
|
Dcb,
|
|
NULL );
|
|
|
|
//
|
|
// Remember this internal open.
|
|
//
|
|
|
|
InterlockedIncrement( (LONG*)&(Dcb->Vcb->InternalOpenCount) );
|
|
|
|
//
|
|
// If this is the root directory, it is also a residual open.
|
|
//
|
|
|
|
if (NodeType( Dcb ) == FAT_NTC_ROOT_DCB) {
|
|
|
|
InterlockedIncrement( (LONG*)&(Dcb->Vcb->ResidualOpenCount) );
|
|
}
|
|
|
|
DirectoryFileObject->SectionObjectPointer = &Dcb->NonPaged->SectionObjectPointers;
|
|
|
|
DirectoryFileObject->ReadAccess = TRUE;
|
|
DirectoryFileObject->WriteAccess = TRUE;
|
|
DirectoryFileObject->DeleteAccess = TRUE;
|
|
|
|
InterlockedIncrement( (LONG*)&Dcb->Specific.Dcb.DirectoryFileOpenCount );
|
|
|
|
Dcb->Specific.Dcb.DirectoryFile = DirectoryFileObject;
|
|
|
|
//
|
|
// Indicate we're happy with the fileobject now.
|
|
//
|
|
|
|
DirectoryFileObject = NULL;
|
|
}
|
|
|
|
} _SEH2_FINALLY {
|
|
|
|
FatReleaseDirectoryFileMutex( Dcb->Vcb );
|
|
|
|
//
|
|
// Rip the object up if we couldn't get the close context.
|
|
//
|
|
|
|
if (DirectoryFileObject) {
|
|
|
|
ObDereferenceObject( DirectoryFileObject );
|
|
}
|
|
} _SEH2_END;
|
|
}
|
|
|
|
//
|
|
// Finally check if we need to initialize the Cache Map for the
|
|
// directory file. The size of the section we are going to map
|
|
// the current allocation size for the directory. Note that the
|
|
// cache manager will provide syncronization for us.
|
|
//
|
|
|
|
if ( Dcb->Specific.Dcb.DirectoryFile->PrivateCacheMap == NULL ) {
|
|
|
|
Dcb->Header.ValidDataLength = FatMaxLarge;
|
|
Dcb->ValidDataToDisk = MAXULONG;
|
|
|
|
FatInitializeCacheMap( Dcb->Specific.Dcb.DirectoryFile,
|
|
(PCC_FILE_SIZES)&Dcb->Header.AllocationSize,
|
|
TRUE,
|
|
&FatData.CacheManagerNoOpCallbacks,
|
|
Dcb );
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatOpenDirectoryFile -> VOID\n", 0);
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
|
|
|
|
PFILE_OBJECT
|
|
FatOpenEaFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PFCB EaFcb
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine opens the Ea file.
|
|
|
|
Arguments:
|
|
|
|
EaFcb - Pointer to the Fcb for the Ea file.
|
|
|
|
Return Value:
|
|
|
|
Pointer to the new file object.
|
|
|
|
--*/
|
|
|
|
{
|
|
PFILE_OBJECT EaFileObject = NULL;
|
|
PDEVICE_OBJECT RealDevice;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatOpenEaFile\n", 0);
|
|
DebugTrace( 0, Dbg, "EaFcb = %p\n", EaFcb);
|
|
|
|
//
|
|
// Create the special file object for the ea file, and set
|
|
// up its pointers back to the Fcb and the section object pointer
|
|
//
|
|
|
|
RealDevice = EaFcb->Vcb->CurrentDevice;
|
|
|
|
EaFileObject = IoCreateStreamFileObject( NULL, RealDevice );
|
|
|
|
_SEH2_TRY {
|
|
|
|
FatPreallocateCloseContext( IrpContext->Vcb);
|
|
|
|
FatSetFileObject( EaFileObject,
|
|
EaFile,
|
|
EaFcb,
|
|
NULL );
|
|
|
|
//
|
|
// Remember this internal, residual open.
|
|
//
|
|
|
|
InterlockedIncrement( (LONG*)&(EaFcb->Vcb->InternalOpenCount) );
|
|
InterlockedIncrement( (LONG*)&(EaFcb->Vcb->ResidualOpenCount) );
|
|
|
|
EaFileObject->SectionObjectPointer = &EaFcb->NonPaged->SectionObjectPointers;
|
|
|
|
EaFileObject->ReadAccess = TRUE;
|
|
EaFileObject->WriteAccess = TRUE;
|
|
|
|
//
|
|
// Finally check if we need to initialize the Cache Map for the
|
|
// ea file. The size of the section we are going to map
|
|
// the current allocation size for the Fcb.
|
|
//
|
|
|
|
EaFcb->Header.ValidDataLength = FatMaxLarge;
|
|
|
|
FatInitializeCacheMap( EaFileObject,
|
|
(PCC_FILE_SIZES)&EaFcb->Header.AllocationSize,
|
|
TRUE,
|
|
&FatData.CacheManagerCallbacks,
|
|
EaFcb );
|
|
|
|
CcSetAdditionalCacheAttributes( EaFileObject, TRUE, TRUE );
|
|
|
|
} _SEH2_FINALLY {
|
|
|
|
//
|
|
// Drop the fileobject if we're raising. Two cases: couldn't get
|
|
// the close context, and it is still an UnopenedFileObject, or
|
|
// we lost trying to build the cache map - in which case we're
|
|
// OK for the close context if we have to.
|
|
//
|
|
|
|
if (_SEH2_AbnormalTermination()) {
|
|
|
|
ObDereferenceObject( EaFileObject );
|
|
}
|
|
} _SEH2_END;
|
|
|
|
DebugTrace(-1, Dbg, "FatOpenEaFile -> %p\n", EaFileObject);
|
|
|
|
UNREFERENCED_PARAMETER( IrpContext );
|
|
|
|
return EaFileObject;
|
|
}
|
|
|
|
|
|
VOID
|
|
FatCloseEaFile (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PVCB Vcb,
|
|
IN BOOLEAN FlushFirst
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine shuts down the ea file. Usually this is required when the volume
|
|
begins to leave the system: after verify, dismount, deletion, pnp.
|
|
|
|
Arguments:
|
|
|
|
Vcb - the volume to close the ea file on
|
|
|
|
FlushFirst - whether the file should be flushed
|
|
|
|
Return Value:
|
|
|
|
None. As a side effect, the EA fileobject in the Vcb is cleared.
|
|
|
|
Caller must have the Vcb exclusive.
|
|
|
|
--*/
|
|
|
|
{
|
|
PFILE_OBJECT EaFileObject = Vcb->VirtualEaFile;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatCloseEaFile\n", 0);
|
|
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb);
|
|
|
|
NT_ASSERT( FatVcbAcquiredExclusive(IrpContext, Vcb) );
|
|
|
|
if (EaFileObject != NULL) {
|
|
|
|
EaFileObject = Vcb->VirtualEaFile;
|
|
|
|
if (FlushFirst) {
|
|
|
|
CcFlushCache( Vcb->VirtualEaFile->SectionObjectPointer, NULL, 0, NULL );
|
|
}
|
|
|
|
Vcb->VirtualEaFile = NULL;
|
|
|
|
//
|
|
// Empty the Mcb for the Ea file.
|
|
//
|
|
|
|
FatRemoveMcbEntry( Vcb, &Vcb->EaFcb->Mcb, 0, 0xFFFFFFFF );
|
|
|
|
//
|
|
// Uninitialize the cache for this file object and dereference it.
|
|
//
|
|
|
|
FatSyncUninitializeCacheMap( IrpContext, EaFileObject );
|
|
|
|
ObDereferenceObject( EaFileObject );
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatCloseEaFile -> %p\n", EaFileObject);
|
|
}
|
|
|
|
|
|
_Requires_lock_held_(_Global_critical_region_)
|
|
VOID
|
|
FatSetDirtyBcb (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PBCB Bcb,
|
|
IN PVCB Vcb OPTIONAL,
|
|
IN BOOLEAN Reversible
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine saves a reference to the bcb in the irp context and
|
|
sets the bcb dirty. This will have the affect of keeping the page in
|
|
memory until we complete the request
|
|
|
|
In addition, a DPC is set to fire in 5 seconds (or if one is pending,
|
|
pushed back 5 seconds) to mark the volume clean.
|
|
|
|
Arguments:
|
|
|
|
Bcb - Supplies the Bcb being set dirty
|
|
|
|
Vcb - Supplies the volume being marked dirty
|
|
|
|
Reversible - Supplies TRUE if the specified range of bcb should be repinned
|
|
so that the changes can be reversed in a controlled fashion if errors
|
|
are encountered.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
|
|
{
|
|
DebugTrace(+1, Dbg, "FatSetDirtyBcb\n", 0 );
|
|
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
|
|
DebugTrace( 0, Dbg, "Bcb = %p\n", Bcb );
|
|
DebugTrace( 0, Dbg, "Vcb = %p\n", Vcb );
|
|
|
|
//
|
|
// Repin the bcb as required
|
|
//
|
|
|
|
if (Reversible) {
|
|
|
|
FatRepinBcb( IrpContext, Bcb );
|
|
}
|
|
|
|
//
|
|
// Set the bcb dirty
|
|
//
|
|
|
|
CcSetDirtyPinnedData( Bcb, NULL );
|
|
|
|
//
|
|
// If volume dirtying isn't disabled for this operation (for
|
|
// instance, when we're changing the dirty state), set the
|
|
// volume dirty if we were given a Vcb that we want to perform
|
|
// clean volume processing on, and return.
|
|
//
|
|
// As a historical note, we used to key off of the old floppy
|
|
// (now deferred flush) bit to disable dirtying behavior. Since
|
|
// hotpluggable media can still be yanked while operations are
|
|
// in flight, recognize that its really the case that FAT12
|
|
// doesn't have the dirty bit.
|
|
//
|
|
|
|
if ( !FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_DIRTY) &&
|
|
ARGUMENT_PRESENT(Vcb) &&
|
|
!FatIsFat12(Vcb)) {
|
|
|
|
KIRQL SavedIrql;
|
|
|
|
BOOLEAN SetTimer;
|
|
|
|
LARGE_INTEGER TimeSincePreviousCall;
|
|
LARGE_INTEGER CurrentTime;
|
|
|
|
//
|
|
// "Borrow" the irp context spinlock.
|
|
//
|
|
|
|
KeQuerySystemTime( &CurrentTime );
|
|
|
|
KeAcquireSpinLock( &FatData.GeneralSpinLock, &SavedIrql );
|
|
|
|
TimeSincePreviousCall.QuadPart =
|
|
CurrentTime.QuadPart - Vcb->LastFatMarkVolumeDirtyCall.QuadPart;
|
|
|
|
//
|
|
// If more than one second has elapsed since the prior call
|
|
// to here, bump the timer up again and see if we need to
|
|
// physically mark the volume dirty.
|
|
//
|
|
|
|
if ( (TimeSincePreviousCall.HighPart != 0) ||
|
|
(TimeSincePreviousCall.LowPart > (1000 * 1000 * 10)) ) {
|
|
|
|
SetTimer = TRUE;
|
|
|
|
} else {
|
|
|
|
SetTimer = FALSE;
|
|
}
|
|
|
|
KeReleaseSpinLock( &FatData.GeneralSpinLock, SavedIrql );
|
|
|
|
if ( SetTimer ) {
|
|
|
|
LARGE_INTEGER CleanVolumeTimer;
|
|
|
|
//
|
|
// We use a shorter volume clean timer for hot plug volumes.
|
|
//
|
|
|
|
CleanVolumeTimer.QuadPart = FlagOn( Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)
|
|
? (LONG)-1500*1000*10
|
|
: (LONG)-8*1000*1000*10;
|
|
|
|
(VOID)KeCancelTimer( &Vcb->CleanVolumeTimer );
|
|
(VOID)KeRemoveQueueDpc( &Vcb->CleanVolumeDpc );
|
|
|
|
//
|
|
// We have now synchronized with anybody clearing the dirty
|
|
// flag, so we can now see if we really have to actually write
|
|
// out the physical bit.
|
|
//
|
|
|
|
if ( !FlagOn(Vcb->VcbState, VCB_STATE_FLAG_VOLUME_DIRTY) ) {
|
|
|
|
//
|
|
// We want to really mark the volume dirty now.
|
|
//
|
|
|
|
if (!FlagOn(Vcb->VcbState, VCB_STATE_FLAG_MOUNTED_DIRTY)) {
|
|
|
|
FatMarkVolume( IrpContext, Vcb, VolumeDirty );
|
|
}
|
|
|
|
SetFlag( Vcb->VcbState, VCB_STATE_FLAG_VOLUME_DIRTY );
|
|
|
|
//
|
|
// Lock the volume if it is removable.
|
|
//
|
|
|
|
if (FlagOn( Vcb->VcbState, VCB_STATE_FLAG_REMOVABLE_MEDIA)) {
|
|
|
|
FatToggleMediaEjectDisable( IrpContext, Vcb, TRUE );
|
|
}
|
|
}
|
|
|
|
KeAcquireSpinLock( &FatData.GeneralSpinLock, &SavedIrql );
|
|
|
|
KeQuerySystemTime( &Vcb->LastFatMarkVolumeDirtyCall );
|
|
|
|
KeReleaseSpinLock( &FatData.GeneralSpinLock, SavedIrql );
|
|
|
|
KeSetTimer( &Vcb->CleanVolumeTimer,
|
|
CleanVolumeTimer,
|
|
&Vcb->CleanVolumeDpc );
|
|
}
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatSetDirtyBcb -> VOID\n", 0 );
|
|
}
|
|
|
|
|
|
VOID
|
|
FatRepinBcb (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PBCB Bcb
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine saves a reference to the bcb in the irp context. This will
|
|
have the affect of keeping the page in memory until we complete the
|
|
request
|
|
|
|
Arguments:
|
|
|
|
Bcb - Supplies the Bcb being referenced
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
|
|
{
|
|
PREPINNED_BCBS Repinned;
|
|
ULONG i;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatRepinBcb\n", 0 );
|
|
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
|
|
DebugTrace( 0, Dbg, "Bcb = %p\n", Bcb );
|
|
|
|
//
|
|
// The algorithm is to search the list of repinned records until
|
|
// we either find a match for the bcb or we find a null slot.
|
|
//
|
|
|
|
Repinned = &IrpContext->Repinned;
|
|
|
|
while (TRUE) {
|
|
|
|
//
|
|
// For every entry in the repinned record check if the bcb's
|
|
// match or if the entry is null. If the bcb's match then
|
|
// we've done because we've already repinned this bcb, if
|
|
// the entry is null then we know, because it's densely packed,
|
|
// that the bcb is not in the list so add it to the repinned
|
|
// record and repin it.
|
|
//
|
|
|
|
for (i = 0; i < REPINNED_BCBS_ARRAY_SIZE; i += 1) {
|
|
|
|
if (Repinned->Bcb[i] == Bcb) {
|
|
|
|
DebugTrace(-1, Dbg, "FatRepinBcb -> VOID\n", 0 );
|
|
return;
|
|
}
|
|
|
|
if (Repinned->Bcb[i] == NULL) {
|
|
|
|
Repinned->Bcb[i] = Bcb;
|
|
CcRepinBcb( Bcb );
|
|
|
|
DebugTrace(-1, Dbg, "FatRepinBcb -> VOID\n", 0 );
|
|
return;
|
|
}
|
|
}
|
|
|
|
//
|
|
// We finished checking one repinned record so now locate the next
|
|
// repinned record, If there isn't one then allocate and zero out
|
|
// a new one.
|
|
//
|
|
|
|
if (Repinned->Next == NULL) {
|
|
|
|
Repinned->Next = FsRtlAllocatePoolWithTag( PagedPool,
|
|
sizeof(REPINNED_BCBS),
|
|
TAG_REPINNED_BCB );
|
|
|
|
RtlZeroMemory( Repinned->Next, sizeof(REPINNED_BCBS) );
|
|
}
|
|
|
|
Repinned = Repinned->Next;
|
|
}
|
|
}
|
|
|
|
|
|
VOID
|
|
FatUnpinRepinnedBcbs (
|
|
IN PIRP_CONTEXT IrpContext
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine frees all of the repinned bcbs, stored in an IRP context.
|
|
|
|
Arguments:
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
|
|
{
|
|
IO_STATUS_BLOCK RaiseIosb;
|
|
PREPINNED_BCBS Repinned;
|
|
BOOLEAN WriteThroughToDisk;
|
|
PFILE_OBJECT FileObject = NULL;
|
|
BOOLEAN ForceVerify = FALSE;
|
|
ULONG i;
|
|
PFCB FcbOrDcb = NULL;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatUnpinRepinnedBcbs\n", 0 );
|
|
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
|
|
|
|
//
|
|
// The algorithm for this procedure is to scan the entire list of
|
|
// repinned records unpinning any repinned bcbs. We start off
|
|
// with the first record in the irp context, and while there is a
|
|
// record to scan we do the following loop.
|
|
//
|
|
|
|
Repinned = &IrpContext->Repinned;
|
|
RaiseIosb.Status = STATUS_SUCCESS;
|
|
|
|
//
|
|
// WinSE bug #307418 "Occasional data corruption when
|
|
// standby/resume while copying files to removable FAT
|
|
// formatted media".
|
|
// Extract main FCB pointer from the irp context - we
|
|
// will need it later to detect new file creation operation.
|
|
//
|
|
|
|
if (IrpContext->MajorFunction == IRP_MJ_CREATE &&
|
|
IrpContext->OriginatingIrp != NULL) {
|
|
PIO_STACK_LOCATION IrpSp;
|
|
|
|
IrpSp = IoGetCurrentIrpStackLocation( IrpContext->OriginatingIrp );
|
|
|
|
if (IrpSp != NULL &&
|
|
IrpSp->FileObject != NULL &&
|
|
IrpSp->FileObject->FsContext != NULL) {
|
|
|
|
FcbOrDcb = IrpSp->FileObject->FsContext;
|
|
}
|
|
}
|
|
|
|
//
|
|
// If the request is write through or the media is deferred flush,
|
|
// unpin the bcb's write through.
|
|
//
|
|
|
|
WriteThroughToDisk = (BOOLEAN) (!FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_WRITE_THROUGH) &&
|
|
IrpContext->Vcb != NULL &&
|
|
(FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WRITE_THROUGH) ||
|
|
FlagOn(IrpContext->Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)));
|
|
|
|
while (Repinned != NULL) {
|
|
|
|
//
|
|
// For every non-null entry in the repinned record unpin the
|
|
// repinned entry.
|
|
//
|
|
// If the this is removable media (therefore all requests write-
|
|
// through) and the write fails, purge the cache so that we throw
|
|
// away the modifications as we will be returning an error to the
|
|
// user.
|
|
//
|
|
|
|
for (i = 0; i < REPINNED_BCBS_ARRAY_SIZE; i += 1) {
|
|
|
|
if (Repinned->Bcb[i] != NULL) {
|
|
|
|
IO_STATUS_BLOCK Iosb;
|
|
|
|
if (WriteThroughToDisk &&
|
|
FlagOn(IrpContext->Vcb->VcbState, VCB_STATE_FLAG_DEFERRED_FLUSH)) {
|
|
|
|
FileObject = CcGetFileObjectFromBcb( Repinned->Bcb[i] );
|
|
}
|
|
|
|
CcUnpinRepinnedBcb( Repinned->Bcb[i],
|
|
WriteThroughToDisk,
|
|
&Iosb );
|
|
|
|
if (!NT_SUCCESS(Iosb.Status)) {
|
|
|
|
if (RaiseIosb.Status == STATUS_SUCCESS) {
|
|
|
|
RaiseIosb = Iosb;
|
|
}
|
|
|
|
//
|
|
// If this was a writethrough device, purge the cache,
|
|
// except for Irp major codes that either don't handle
|
|
// the error paths correctly or are simple victims like
|
|
// cleanup.c.
|
|
//
|
|
|
|
if (FileObject &&
|
|
(IrpContext->MajorFunction != IRP_MJ_CLEANUP) &&
|
|
(IrpContext->MajorFunction != IRP_MJ_FLUSH_BUFFERS) &&
|
|
(IrpContext->MajorFunction != IRP_MJ_SET_INFORMATION)
|
|
|
|
&&
|
|
|
|
//
|
|
// WinSE bug #307418 "Occasional data corruption when
|
|
// standby/resume while copying files to removable FAT
|
|
// formatted media".
|
|
// Buffer unpinning for new file creation operation can
|
|
// be interrupted by system syspend. As a result some BCBs
|
|
// will be successfully written to the disk while others will
|
|
// be kicked back with STATUS_VERIFY_REQUIRED. Since there is
|
|
// is still a chance for the failed BCBs to reach the disk
|
|
// after the volume verification we'll not purge them.
|
|
// Instead FatCommonCreate() will unroll the file creation
|
|
// changes for these pages.
|
|
//
|
|
|
|
!(IrpContext->MajorFunction == IRP_MJ_CREATE &&
|
|
Iosb.Status == STATUS_VERIFY_REQUIRED &&
|
|
FcbOrDcb != NULL &&
|
|
NodeType( FcbOrDcb ) == FAT_NTC_FCB)) {
|
|
|
|
//
|
|
// The call to CcPurgeCacheSection() below will
|
|
// purge the entire file from memory. It will also
|
|
// block until all the file's BCB's are pinned.
|
|
//
|
|
// We end up in a deadlock situation of there
|
|
// are any other pinned BCB's in this IRP context
|
|
// so the first thing we do is search the list
|
|
// for BCB's pinned in the same file and unpin
|
|
// them.
|
|
//
|
|
// We are probably not going to lose data because
|
|
// it's safe to assume that all flushes will
|
|
// fail after the first one fails.
|
|
//
|
|
|
|
ULONG j;
|
|
ULONG k = i + 1;
|
|
PREPINNED_BCBS RepinnedToPurge = Repinned;
|
|
|
|
while( RepinnedToPurge != NULL ) {
|
|
|
|
for (j = k; j < REPINNED_BCBS_ARRAY_SIZE; j++) {
|
|
|
|
if (RepinnedToPurge->Bcb[j] != NULL) {
|
|
|
|
if (CcGetFileObjectFromBcb( RepinnedToPurge->Bcb[j] ) == FileObject) {
|
|
|
|
CcUnpinRepinnedBcb( RepinnedToPurge->Bcb[j],
|
|
FALSE,
|
|
&Iosb );
|
|
|
|
RepinnedToPurge->Bcb[j] = NULL;
|
|
}
|
|
}
|
|
}
|
|
|
|
RepinnedToPurge = RepinnedToPurge->Next;
|
|
k = 0;
|
|
}
|
|
|
|
CcPurgeCacheSection( FileObject->SectionObjectPointer,
|
|
NULL,
|
|
0,
|
|
FALSE );
|
|
|
|
//
|
|
// Force a verify operation here since who knows
|
|
// what state things are in.
|
|
//
|
|
|
|
ForceVerify = TRUE;
|
|
}
|
|
}
|
|
|
|
Repinned->Bcb[i] = NULL;
|
|
|
|
}
|
|
}
|
|
|
|
//
|
|
// Now find the next repinned record in the list, and possibly
|
|
// delete the one we've just processed.
|
|
//
|
|
|
|
if (Repinned != &IrpContext->Repinned) {
|
|
|
|
PREPINNED_BCBS Saved;
|
|
|
|
Saved = Repinned->Next;
|
|
ExFreePool( Repinned );
|
|
Repinned = Saved;
|
|
|
|
} else {
|
|
|
|
Repinned = Repinned->Next;
|
|
IrpContext->Repinned.Next = NULL;
|
|
}
|
|
}
|
|
|
|
//
|
|
// Now if we weren't completely successful in the our unpin
|
|
// then raise the iosb we got
|
|
//
|
|
|
|
if (!NT_SUCCESS(RaiseIosb.Status)) {
|
|
|
|
if (ForceVerify && FileObject) {
|
|
|
|
SetFlag(FileObject->DeviceObject->Flags, DO_VERIFY_VOLUME);
|
|
|
|
IoSetHardErrorOrVerifyDevice( IrpContext->OriginatingIrp,
|
|
FileObject->DeviceObject );
|
|
}
|
|
|
|
if (!FlagOn( IrpContext->Flags, IRP_CONTEXT_FLAG_DISABLE_RAISE )) {
|
|
if (IrpContext->OriginatingIrp) {
|
|
IrpContext->OriginatingIrp->IoStatus = RaiseIosb;
|
|
}
|
|
FatNormalizeAndRaiseStatus( IrpContext, RaiseIosb.Status );
|
|
}
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatUnpinRepinnedBcbs -> VOID\n", 0 );
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
FINISHED
|
|
FatZeroData (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PVCB Vcb,
|
|
IN PFILE_OBJECT FileObject,
|
|
IN ULONG StartingZero,
|
|
IN ULONG ByteCount
|
|
)
|
|
|
|
/*++
|
|
|
|
**** Temporary function - Remove when CcZeroData is capable of handling
|
|
non sector aligned requests.
|
|
|
|
--*/
|
|
{
|
|
#ifndef __REACTOS__
|
|
LARGE_INTEGER ZeroStart = {0,0};
|
|
LARGE_INTEGER BeyondZeroEnd = {0,0};
|
|
#else
|
|
LARGE_INTEGER ZeroStart = {{0,0}};
|
|
LARGE_INTEGER BeyondZeroEnd = {{0,0}};
|
|
#endif
|
|
|
|
ULONG SectorSize;
|
|
|
|
BOOLEAN Finished;
|
|
|
|
PAGED_CODE();
|
|
|
|
SectorSize = (ULONG)Vcb->Bpb.BytesPerSector;
|
|
|
|
ZeroStart.LowPart = (StartingZero + (SectorSize - 1)) & ~(SectorSize - 1);
|
|
|
|
//
|
|
// Detect overflow if we were asked to zero in the last sector of the file,
|
|
// which must be "zeroed" already (or we're in trouble).
|
|
//
|
|
|
|
if (StartingZero != 0 && ZeroStart.LowPart == 0) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
//
|
|
// Note that BeyondZeroEnd can take the value 4gb.
|
|
//
|
|
|
|
BeyondZeroEnd.QuadPart = ((ULONGLONG) StartingZero + ByteCount + (SectorSize - 1))
|
|
& (~((LONGLONG) SectorSize - 1));
|
|
|
|
//
|
|
// If we were called to just zero part of a sector we are in trouble.
|
|
//
|
|
|
|
if ( ZeroStart.QuadPart == BeyondZeroEnd.QuadPart ) {
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
Finished = CcZeroData( FileObject,
|
|
&ZeroStart,
|
|
&BeyondZeroEnd,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT) );
|
|
|
|
return Finished;
|
|
}
|
|
|
|
|
|
NTSTATUS
|
|
FatCompleteMdl (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PIRP Irp
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine performs the function of completing Mdl read and write
|
|
requests. It should be called only from FatFsdRead and FatFsdWrite.
|
|
|
|
Arguments:
|
|
|
|
Irp - Supplies the originating Irp.
|
|
|
|
Return Value:
|
|
|
|
NTSTATUS - Will always be STATUS_PENDING or STATUS_SUCCESS.
|
|
|
|
--*/
|
|
|
|
{
|
|
PFILE_OBJECT FileObject;
|
|
PIO_STACK_LOCATION IrpSp;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatCompleteMdl\n", 0 );
|
|
DebugTrace( 0, Dbg, "IrpContext = %p\n", IrpContext );
|
|
DebugTrace( 0, Dbg, "Irp = %p\n", Irp );
|
|
|
|
//
|
|
// Do completion processing.
|
|
//
|
|
|
|
FileObject = IoGetCurrentIrpStackLocation( Irp )->FileObject;
|
|
|
|
switch( IrpContext->MajorFunction ) {
|
|
|
|
case IRP_MJ_READ:
|
|
|
|
CcMdlReadComplete( FileObject, Irp->MdlAddress );
|
|
break;
|
|
|
|
case IRP_MJ_WRITE:
|
|
|
|
IrpSp = IoGetCurrentIrpStackLocation( Irp );
|
|
|
|
NT_ASSERT( FlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT ));
|
|
|
|
CcMdlWriteComplete( FileObject, &IrpSp->Parameters.Write.ByteOffset, Irp->MdlAddress );
|
|
|
|
Irp->IoStatus.Status = STATUS_SUCCESS;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
DebugTrace( DEBUG_TRACE_ERROR, 0, "Illegal Mdl Complete.\n", 0);
|
|
#ifdef _MSC_VER
|
|
#pragma prefast( suppress: 28159, "we're very broken if we get here" )
|
|
#endif
|
|
FatBugCheck( IrpContext->MajorFunction, 0, 0 );
|
|
}
|
|
|
|
//
|
|
// Mdl is now deallocated.
|
|
//
|
|
|
|
Irp->MdlAddress = NULL;
|
|
|
|
//
|
|
// Complete the request and exit right away.
|
|
//
|
|
|
|
FatCompleteRequest( IrpContext, Irp, STATUS_SUCCESS );
|
|
|
|
DebugTrace(-1, Dbg, "FatCompleteMdl -> STATUS_SUCCESS\n", 0 );
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
VOID
|
|
FatSyncUninitializeCacheMap (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PFILE_OBJECT FileObject
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
The routine performs a CcUnitializeCacheMap to LargeZero synchronously. That
|
|
is it waits on the Cc event. This call is useful when we want to be certain
|
|
when a close will actually some in.
|
|
|
|
Return Value:
|
|
|
|
None.
|
|
|
|
--*/
|
|
|
|
{
|
|
CACHE_UNINITIALIZE_EVENT UninitializeCompleteEvent;
|
|
NTSTATUS WaitStatus;
|
|
|
|
UNREFERENCED_PARAMETER( IrpContext );
|
|
|
|
PAGED_CODE();
|
|
|
|
KeInitializeEvent( &UninitializeCompleteEvent.Event,
|
|
SynchronizationEvent,
|
|
FALSE);
|
|
|
|
CcUninitializeCacheMap( FileObject,
|
|
&FatLargeZero,
|
|
&UninitializeCompleteEvent );
|
|
|
|
//
|
|
// Now wait for the cache manager to finish purging the file.
|
|
// This will garentee that Mm gets the purge before we
|
|
// delete the Vcb.
|
|
//
|
|
|
|
#ifdef _MSC_VER
|
|
#pragma prefast( suppress: 28931, "we use WaitStatus in the debug assert, in fre builds prefast complains it's unused" )
|
|
#endif
|
|
WaitStatus = KeWaitForSingleObject( &UninitializeCompleteEvent.Event,
|
|
Executive,
|
|
KernelMode,
|
|
FALSE,
|
|
NULL);
|
|
|
|
NT_ASSERT(WaitStatus == STATUS_SUCCESS);
|
|
}
|
|
|
|
VOID
|
|
FatPinMappedData (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PDCB Dcb,
|
|
IN VBO StartingVbo,
|
|
IN ULONG ByteCount,
|
|
OUT PBCB *Bcb
|
|
)
|
|
|
|
/*++
|
|
|
|
Routine Description:
|
|
|
|
This routine pins data that was previously mapped before setting it dirty.
|
|
|
|
Arguments:
|
|
|
|
Dcb - Pointer to the DCB for the directory
|
|
|
|
StartingVbo - The virtual offset of the first desired byte
|
|
|
|
ByteCount - Number of bytes desired
|
|
|
|
Bcb - Returns a pointer to the BCB which is valid until unpinned
|
|
|
|
--*/
|
|
|
|
{
|
|
LARGE_INTEGER Vbo;
|
|
|
|
PAGED_CODE();
|
|
|
|
DebugTrace(+1, Dbg, "FatPinMappedData\n", 0);
|
|
DebugTrace( 0, Dbg, "Dcb = %p\n", Dcb);
|
|
DebugTrace( 0, Dbg, "StartingVbo = %08lx\n", StartingVbo);
|
|
DebugTrace( 0, Dbg, "ByteCount = %08lx\n", ByteCount);
|
|
|
|
//
|
|
// Call the Cache manager to perform the operation.
|
|
//
|
|
|
|
Vbo.QuadPart = StartingVbo;
|
|
|
|
if (!CcPinMappedData( Dcb->Specific.Dcb.DirectoryFile,
|
|
&Vbo,
|
|
ByteCount,
|
|
BooleanFlagOn(IrpContext->Flags, IRP_CONTEXT_FLAG_WAIT),
|
|
Bcb )) {
|
|
|
|
//
|
|
// Could not pin the data without waiting (cache miss).
|
|
//
|
|
|
|
FatRaiseStatus( IrpContext, STATUS_CANT_WAIT );
|
|
}
|
|
|
|
DebugTrace(-1, Dbg, "FatReadDirectoryFile -> VOID, *BCB = %p\n", *Bcb);
|
|
|
|
return;
|
|
}
|
|
|
|
#if (NTDDI_VERSION >= NTDDI_WIN8)
|
|
|
|
NTSTATUS
|
|
FatPrefetchPages (
|
|
IN PIRP_CONTEXT IrpContext,
|
|
IN PFILE_OBJECT FileObject,
|
|
IN ULONG StartingPage,
|
|
IN ULONG PageCount
|
|
)
|
|
{
|
|
IO_PRIORITY_INFO PriorityInformation = {0};
|
|
MM_PREFETCH_FLAGS PrefetchFlags;
|
|
ULONG PageNo;
|
|
NTSTATUS Status;
|
|
|
|
PREAD_LIST ReadList = NULL;
|
|
|
|
UNREFERENCED_PARAMETER( IrpContext );
|
|
|
|
PAGED_CODE();
|
|
|
|
//
|
|
// Succeed zero page prefetch requests.
|
|
//
|
|
|
|
if (PageCount == 0) {
|
|
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
//
|
|
// Mm's prefetch API's "only" support fetching a ULONG worth of pages.
|
|
// Make sure we don't overflow.
|
|
//
|
|
|
|
ASSERT( PageCount < (PFN_NUMBER)MAXULONG );
|
|
|
|
IoInitializePriorityInfo( &PriorityInformation );
|
|
|
|
Status = IoRetrievePriorityInfo( IrpContext->OriginatingIrp,
|
|
FileObject,
|
|
IrpContext->OriginatingIrp->Tail.Overlay.Thread,
|
|
&PriorityInformation );
|
|
|
|
if (!NT_SUCCESS( Status)) {
|
|
|
|
goto Cleanup;
|
|
}
|
|
|
|
ReadList = ExAllocatePoolWithTag( PagedPool,
|
|
FIELD_OFFSET( READ_LIST, List ) + PageCount * sizeof( FILE_SEGMENT_ELEMENT ),
|
|
' taF' );
|
|
|
|
if (ReadList == NULL) {
|
|
|
|
Status = STATUS_INSUFFICIENT_RESOURCES;
|
|
goto Cleanup;
|
|
}
|
|
|
|
//
|
|
// Call Mm to prefetch data.
|
|
//
|
|
|
|
ReadList->FileObject = FileObject;
|
|
ReadList->IsImage = FALSE;
|
|
ReadList->NumberOfEntries = PageCount;
|
|
|
|
PrefetchFlags.AllFlags = 0;
|
|
PrefetchFlags.Flags.Priority = PriorityInformation.PagePriority;
|
|
PrefetchFlags.Flags.RepurposePriority = SYSTEM_PAGE_PRIORITY_LEVELS - 1;
|
|
PrefetchFlags.Flags.PriorityProtection = 1;
|
|
ReadList->List[0].Alignment = StartingPage * PAGE_SIZE;
|
|
ReadList->List[0].Alignment |= PrefetchFlags.AllFlags;
|
|
|
|
for (PageNo = 1; PageNo < PageCount; PageNo++) {
|
|
|
|
ReadList->List[PageNo].Alignment = ReadList->List[PageNo-1].Alignment + PAGE_SIZE;
|
|
}
|
|
|
|
Status = MmPrefetchPages( 1, &ReadList );
|
|
|
|
Cleanup:
|
|
|
|
if (ReadList != NULL) {
|
|
|
|
ExFreePoolWithTag( ReadList, ' taF' );
|
|
}
|
|
|
|
return Status;
|
|
}
|
|
#endif
|
|
|