mirror of
https://github.com/reactos/reactos.git
synced 2024-10-30 11:35:58 +00:00
a3c58ca76f
The reserve IRP is an IRP which is allocated on system boot and kept during the whole system life. Its purpose is to allow page reads in case of low-memory situations where the system doesn't have enough memory left to allocate an IRP to read from the page file (would be catastrophic situation).
2016 lines
57 KiB
C
2016 lines
57 KiB
C
/*
|
|
* PROJECT: ReactOS Kernel
|
|
* LICENSE: GPL - See COPYING in the top level directory
|
|
* FILE: ntoskrnl/io/iomgr/irp.c
|
|
* PURPOSE: IRP Handling Functions
|
|
* PROGRAMMERS: Alex Ionescu (alex.ionescu@reactos.org)
|
|
* Gunnar Dalsnes
|
|
* Filip Navara (navaraf@reactos.org)
|
|
* Pierre Schweitzer (pierre@reactos.org)
|
|
*/
|
|
|
|
/* INCLUDES ****************************************************************/
|
|
|
|
#include <ntoskrnl.h>
|
|
#define NDEBUG
|
|
#include <debug.h>
|
|
|
|
PIRP IopDeadIrp;
|
|
RESERVE_IRP_ALLOCATOR IopReserveIrpAllocator;
|
|
|
|
/* PRIVATE FUNCTIONS ********************************************************/
|
|
|
|
VOID
|
|
NTAPI
|
|
IopFreeIrpKernelApc(IN PKAPC Apc,
|
|
IN PKNORMAL_ROUTINE *NormalRoutine,
|
|
IN PVOID *NormalContext,
|
|
IN PVOID *SystemArgument1,
|
|
IN PVOID *SystemArgument2)
|
|
{
|
|
/* Free the IRP */
|
|
IoFreeIrp(CONTAINING_RECORD(Apc, IRP, Tail.Apc));
|
|
}
|
|
|
|
VOID
|
|
NTAPI
|
|
IopAbortIrpKernelApc(IN PKAPC Apc)
|
|
{
|
|
/* Free the IRP */
|
|
IoFreeIrp(CONTAINING_RECORD(Apc, IRP, Tail.Apc));
|
|
}
|
|
|
|
NTSTATUS
|
|
NTAPI
|
|
IopCleanupFailedIrp(IN PFILE_OBJECT FileObject,
|
|
IN PKEVENT EventObject OPTIONAL,
|
|
IN PVOID Buffer OPTIONAL)
|
|
{
|
|
PAGED_CODE();
|
|
|
|
/* Dereference the event */
|
|
if (EventObject) ObDereferenceObject(EventObject);
|
|
|
|
/* Free a buffer, if any */
|
|
if (Buffer) ExFreePool(Buffer);
|
|
|
|
/* If this was a file opened for synch I/O, then unlock it */
|
|
if (FileObject->Flags & FO_SYNCHRONOUS_IO) IopUnlockFileObject(FileObject);
|
|
|
|
/* Now dereference it and return */
|
|
ObDereferenceObject(FileObject);
|
|
return STATUS_INSUFFICIENT_RESOURCES;
|
|
}
|
|
|
|
VOID
|
|
NTAPI
|
|
IopAbortInterruptedIrp(IN PKEVENT EventObject,
|
|
IN PIRP Irp)
|
|
{
|
|
KIRQL OldIrql;
|
|
BOOLEAN CancelResult;
|
|
LARGE_INTEGER Wait;
|
|
PAGED_CODE();
|
|
|
|
/* Raise IRQL to APC */
|
|
KeRaiseIrql(APC_LEVEL, &OldIrql);
|
|
|
|
/* Check if nobody completed it yet */
|
|
if (!KeReadStateEvent(EventObject))
|
|
{
|
|
/* First, cancel it */
|
|
CancelResult = IoCancelIrp(Irp);
|
|
KeLowerIrql(OldIrql);
|
|
|
|
/* Check if we cancelled it */
|
|
if (CancelResult)
|
|
{
|
|
/* Wait for the IRP to be cancelled */
|
|
Wait.QuadPart = -100000;
|
|
while (!KeReadStateEvent(EventObject))
|
|
{
|
|
/* Delay indefintely */
|
|
KeDelayExecutionThread(KernelMode, FALSE, &Wait);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* No cancellation done, so wait for the I/O system to kill it */
|
|
KeWaitForSingleObject(EventObject,
|
|
Executive,
|
|
KernelMode,
|
|
FALSE,
|
|
NULL);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/* We got preempted, so give up */
|
|
KeLowerIrql(OldIrql);
|
|
}
|
|
}
|
|
|
|
VOID
|
|
NTAPI
|
|
IopDisassociateThreadIrp(VOID)
|
|
{
|
|
KIRQL OldIrql, LockIrql;
|
|
PETHREAD IrpThread;
|
|
PLIST_ENTRY IrpEntry;
|
|
PIO_ERROR_LOG_PACKET ErrorLogEntry;
|
|
PDEVICE_OBJECT DeviceObject = NULL;
|
|
PIO_STACK_LOCATION IoStackLocation;
|
|
|
|
/* First, raise to APC to protect IrpList */
|
|
KeRaiseIrql(APC_LEVEL, &OldIrql);
|
|
|
|
/* Get the Thread and check the list */
|
|
IrpThread = PsGetCurrentThread();
|
|
if (IsListEmpty(&IrpThread->IrpList))
|
|
{
|
|
/* It got completed now, so quit */
|
|
KeLowerIrql(OldIrql);
|
|
return;
|
|
}
|
|
|
|
/* Ensure no one will come disturb */
|
|
LockIrql = KeAcquireQueuedSpinLock(LockQueueIoCompletionLock);
|
|
|
|
/* Get the misbehaving IRP */
|
|
IrpEntry = IrpThread->IrpList.Flink;
|
|
IopDeadIrp = CONTAINING_RECORD(IrpEntry, IRP, ThreadListEntry);
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Deassociating IRP %p for %p\n",
|
|
__FUNCTION__,
|
|
IopDeadIrp,
|
|
IrpThread);
|
|
|
|
/* Don't cancel the IRP if it's already been completed far */
|
|
if (IopDeadIrp->CurrentLocation == (IopDeadIrp->StackCount + 2))
|
|
{
|
|
/* Return */
|
|
KeReleaseQueuedSpinLock(LockQueueIoCompletionLock, LockIrql);
|
|
KeLowerIrql(OldIrql);
|
|
return;
|
|
}
|
|
|
|
/* Disown the IRP! */
|
|
IopDeadIrp->Tail.Overlay.Thread = NULL;
|
|
RemoveHeadList(&IrpThread->IrpList);
|
|
InitializeListHead(&IopDeadIrp->ThreadListEntry);
|
|
|
|
/* Get the stack location and check if it's valid */
|
|
IoStackLocation = IoGetCurrentIrpStackLocation(IopDeadIrp);
|
|
if (IopDeadIrp->CurrentLocation <= IopDeadIrp->StackCount)
|
|
{
|
|
/* Get the device object */
|
|
DeviceObject = IoStackLocation->DeviceObject;
|
|
}
|
|
|
|
KeReleaseQueuedSpinLock(LockQueueIoCompletionLock, LockIrql);
|
|
/* Lower IRQL now, since we have the pointers we need */
|
|
KeLowerIrql(OldIrql);
|
|
|
|
/* Check if we can send an Error Log Entry*/
|
|
if (DeviceObject)
|
|
{
|
|
/* Allocate an entry */
|
|
ErrorLogEntry = IoAllocateErrorLogEntry(DeviceObject,
|
|
sizeof(IO_ERROR_LOG_PACKET));
|
|
if (ErrorLogEntry)
|
|
{
|
|
/* Write the entry */
|
|
ErrorLogEntry->ErrorCode = IO_DRIVER_CANCEL_TIMEOUT;
|
|
IoWriteErrorLogEntry(ErrorLogEntry);
|
|
}
|
|
}
|
|
}
|
|
|
|
VOID
|
|
NTAPI
|
|
IopCleanupIrp(IN PIRP Irp,
|
|
IN PFILE_OBJECT FileObject)
|
|
{
|
|
PMDL Mdl;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Cleaning IRP %p for %p\n",
|
|
__FUNCTION__,
|
|
Irp,
|
|
FileObject);
|
|
|
|
/* Check if there's an MDL */
|
|
while ((Mdl = Irp->MdlAddress))
|
|
{
|
|
/* Clear all of them */
|
|
Irp->MdlAddress = Mdl->Next;
|
|
IoFreeMdl(Mdl);
|
|
}
|
|
|
|
/* Check if the IRP has system buffer */
|
|
if (Irp->Flags & IRP_DEALLOCATE_BUFFER)
|
|
{
|
|
/* Free the buffer */
|
|
ExFreePoolWithTag(Irp->AssociatedIrp.SystemBuffer, TAG_SYS_BUF);
|
|
}
|
|
|
|
/* Check if this IRP has a user event, a file object, and is async */
|
|
if ((Irp->UserEvent) &&
|
|
!(Irp->Flags & IRP_SYNCHRONOUS_API) &&
|
|
(FileObject))
|
|
{
|
|
/* Dereference the User Event */
|
|
ObDereferenceObject(Irp->UserEvent);
|
|
}
|
|
|
|
/* Check if we have a file object and this isn't a create operation */
|
|
if ((FileObject) && !(Irp->Flags & IRP_CREATE_OPERATION))
|
|
{
|
|
/* Dereference the file object */
|
|
ObDereferenceObject(FileObject);
|
|
}
|
|
|
|
/* Free the IRP */
|
|
IoFreeIrp(Irp);
|
|
}
|
|
|
|
VOID
|
|
NTAPI
|
|
IopCompleteRequest(IN PKAPC Apc,
|
|
IN PKNORMAL_ROUTINE* NormalRoutine,
|
|
IN PVOID* NormalContext,
|
|
IN PVOID* SystemArgument1,
|
|
IN PVOID* SystemArgument2)
|
|
{
|
|
PFILE_OBJECT FileObject;
|
|
PIRP Irp;
|
|
PMDL Mdl, NextMdl;
|
|
PVOID Port = NULL, Key = NULL;
|
|
BOOLEAN SignaledCreateRequest = FALSE;
|
|
|
|
/* Get data from the APC */
|
|
FileObject = (PFILE_OBJECT)*SystemArgument1;
|
|
Irp = CONTAINING_RECORD(Apc, IRP, Tail.Apc);
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Completing IRP %p for %p\n",
|
|
__FUNCTION__,
|
|
Irp,
|
|
FileObject);
|
|
|
|
/* Sanity check */
|
|
ASSERT(Irp->IoStatus.Status != (NTSTATUS)0xFFFFFFFF);
|
|
|
|
/* Check if we have a file object */
|
|
if (*SystemArgument2)
|
|
{
|
|
/* Check if we're reparsing */
|
|
if ((Irp->IoStatus.Status == STATUS_REPARSE) &&
|
|
(Irp->IoStatus.Information == IO_REPARSE_TAG_MOUNT_POINT))
|
|
{
|
|
PREPARSE_DATA_BUFFER ReparseData;
|
|
|
|
ReparseData = (PREPARSE_DATA_BUFFER)*SystemArgument2;
|
|
|
|
ASSERT(ReparseData->ReparseTag == IO_REPARSE_TAG_MOUNT_POINT);
|
|
ASSERT(ReparseData->ReparseDataLength < MAXIMUM_REPARSE_DATA_BUFFER_SIZE);
|
|
ASSERT(ReparseData->Reserved < MAXIMUM_REPARSE_DATA_BUFFER_SIZE);
|
|
|
|
IopDoNameTransmogrify(Irp, FileObject, ReparseData);
|
|
}
|
|
}
|
|
|
|
/* Handle Buffered case first */
|
|
if (Irp->Flags & IRP_BUFFERED_IO)
|
|
{
|
|
/* Check if we have an input buffer and if we succeeded */
|
|
if ((Irp->Flags & IRP_INPUT_OPERATION) &&
|
|
(Irp->IoStatus.Status != STATUS_VERIFY_REQUIRED) &&
|
|
!(NT_ERROR(Irp->IoStatus.Status)))
|
|
{
|
|
_SEH2_TRY
|
|
{
|
|
/* Copy the buffer back to the user */
|
|
RtlCopyMemory(Irp->UserBuffer,
|
|
Irp->AssociatedIrp.SystemBuffer,
|
|
Irp->IoStatus.Information);
|
|
}
|
|
_SEH2_EXCEPT(EXCEPTION_EXECUTE_HANDLER)
|
|
{
|
|
/* Fail the IRP */
|
|
Irp->IoStatus.Status = _SEH2_GetExceptionCode();
|
|
}
|
|
_SEH2_END;
|
|
}
|
|
|
|
/* Also check if we should de-allocate it */
|
|
if (Irp->Flags & IRP_DEALLOCATE_BUFFER)
|
|
{
|
|
/* Deallocate it */
|
|
ExFreePool(Irp->AssociatedIrp.SystemBuffer);
|
|
}
|
|
}
|
|
|
|
/* Now we got rid of these two... */
|
|
Irp->Flags &= ~(IRP_BUFFERED_IO | IRP_DEALLOCATE_BUFFER);
|
|
|
|
/* Check if there's an MDL */
|
|
for (Mdl = Irp->MdlAddress; Mdl; Mdl = NextMdl)
|
|
{
|
|
/* Free it */
|
|
NextMdl = Mdl->Next;
|
|
IoFreeMdl(Mdl);
|
|
}
|
|
|
|
/* No MDLs left */
|
|
Irp->MdlAddress = NULL;
|
|
|
|
/*
|
|
* Check if either the request was completed without any errors
|
|
* (but warnings are OK!), or if it was completed with an error, but
|
|
* did return from a pending I/O Operation and is not synchronous.
|
|
*/
|
|
if (!NT_ERROR(Irp->IoStatus.Status) ||
|
|
(Irp->PendingReturned &&
|
|
!IsIrpSynchronous(Irp, FileObject)))
|
|
{
|
|
/* Get any information we need from the FO before we kill it */
|
|
if ((FileObject) && (FileObject->CompletionContext))
|
|
{
|
|
/* Save Completion Data */
|
|
Port = FileObject->CompletionContext->Port;
|
|
Key = FileObject->CompletionContext->Key;
|
|
}
|
|
|
|
/* Check for UserIos */
|
|
if (Irp->UserIosb != NULL)
|
|
{
|
|
/* Use SEH to make sure we don't write somewhere invalid */
|
|
_SEH2_TRY
|
|
{
|
|
/* Save the IOSB Information */
|
|
*Irp->UserIosb = Irp->IoStatus;
|
|
}
|
|
_SEH2_EXCEPT(EXCEPTION_EXECUTE_HANDLER)
|
|
{
|
|
/* Ignore any error */
|
|
}
|
|
_SEH2_END;
|
|
}
|
|
|
|
/* Check if we have an event or a file object */
|
|
if (Irp->UserEvent)
|
|
{
|
|
/* At the very least, this is a PKEVENT, so signal it always */
|
|
KeSetEvent(Irp->UserEvent, 0, FALSE);
|
|
|
|
/* Check if we also have a File Object */
|
|
if (FileObject)
|
|
{
|
|
/* Check if this is an Asynch API */
|
|
if (!(Irp->Flags & IRP_SYNCHRONOUS_API))
|
|
{
|
|
/* Dereference the event */
|
|
ObDereferenceObject(Irp->UserEvent);
|
|
}
|
|
|
|
/*
|
|
* Now, if this is a Synch I/O File Object, then this event is
|
|
* NOT an actual Executive Event, so we won't dereference it,
|
|
* and instead, we will signal the File Object
|
|
*/
|
|
if ((FileObject->Flags & FO_SYNCHRONOUS_IO) &&
|
|
!(Irp->Flags & IRP_OB_QUERY_NAME))
|
|
{
|
|
/* Signal the file object and set the status */
|
|
KeSetEvent(&FileObject->Event, 0, FALSE);
|
|
FileObject->FinalStatus = Irp->IoStatus.Status;
|
|
}
|
|
|
|
/*
|
|
* This could also be a create operation, in which case we want
|
|
* to make sure there's no APC fired.
|
|
*/
|
|
if (Irp->Flags & IRP_CREATE_OPERATION)
|
|
{
|
|
/* Clear the APC Routine and remember this */
|
|
Irp->Overlay.AsynchronousParameters.UserApcRoutine = NULL;
|
|
SignaledCreateRequest = TRUE;
|
|
}
|
|
}
|
|
}
|
|
else if (FileObject)
|
|
{
|
|
/* Signal the file object and set the status */
|
|
KeSetEvent(&FileObject->Event, 0, FALSE);
|
|
FileObject->FinalStatus = Irp->IoStatus.Status;
|
|
|
|
/*
|
|
* This could also be a create operation, in which case we want
|
|
* to make sure there's no APC fired.
|
|
*/
|
|
if (Irp->Flags & IRP_CREATE_OPERATION)
|
|
{
|
|
/* Clear the APC Routine and remember this */
|
|
Irp->Overlay.AsynchronousParameters.UserApcRoutine = NULL;
|
|
SignaledCreateRequest = TRUE;
|
|
}
|
|
}
|
|
|
|
/* Update transfer count for everything but create operation */
|
|
if (!(Irp->Flags & IRP_CREATE_OPERATION))
|
|
{
|
|
if (Irp->Flags & IRP_WRITE_OPERATION)
|
|
{
|
|
/* Update write transfer count */
|
|
IopUpdateTransferCount(IopWriteTransfer,
|
|
(ULONG)Irp->IoStatus.Information);
|
|
}
|
|
else if (Irp->Flags & IRP_READ_OPERATION)
|
|
{
|
|
/* Update read transfer count */
|
|
IopUpdateTransferCount(IopReadTransfer,
|
|
(ULONG)Irp->IoStatus.Information);
|
|
}
|
|
else
|
|
{
|
|
/* Update other transfer count */
|
|
IopUpdateTransferCount(IopOtherTransfer,
|
|
(ULONG)Irp->IoStatus.Information);
|
|
}
|
|
}
|
|
|
|
/* Now that we've signaled the events, de-associate the IRP */
|
|
IopUnQueueIrpFromThread(Irp);
|
|
|
|
/* Now check if a User APC Routine was requested */
|
|
if (Irp->Overlay.AsynchronousParameters.UserApcRoutine)
|
|
{
|
|
/* Initialize it */
|
|
KeInitializeApc(&Irp->Tail.Apc,
|
|
KeGetCurrentThread(),
|
|
CurrentApcEnvironment,
|
|
IopFreeIrpKernelApc,
|
|
IopAbortIrpKernelApc,
|
|
(PKNORMAL_ROUTINE)Irp->
|
|
Overlay.AsynchronousParameters.UserApcRoutine,
|
|
Irp->RequestorMode,
|
|
Irp->
|
|
Overlay.AsynchronousParameters.UserApcContext);
|
|
|
|
/* Queue it */
|
|
KeInsertQueueApc(&Irp->Tail.Apc, Irp->UserIosb, NULL, 2);
|
|
}
|
|
else if ((Port) &&
|
|
(Irp->Overlay.AsynchronousParameters.UserApcContext))
|
|
{
|
|
/* We have an I/O Completion setup... create the special Overlay */
|
|
Irp->Tail.CompletionKey = Key;
|
|
Irp->Tail.Overlay.PacketType = IopCompletionPacketIrp;
|
|
KeInsertQueue(Port, &Irp->Tail.Overlay.ListEntry);
|
|
}
|
|
else
|
|
{
|
|
/* Free the IRP since we don't need it anymore */
|
|
IoFreeIrp(Irp);
|
|
}
|
|
|
|
/* Check if we have a file object that wasn't part of a create */
|
|
if ((FileObject) && !(SignaledCreateRequest))
|
|
{
|
|
/* Dereference it, since it's not needed anymore either */
|
|
ObDereferenceObjectDeferDelete(FileObject);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* Either we didn't return from the request, or we did return but this
|
|
* request was synchronous.
|
|
*/
|
|
if ((Irp->PendingReturned) && (FileObject))
|
|
{
|
|
/* So we did return with a synch operation, was it the IRP? */
|
|
if (Irp->Flags & IRP_SYNCHRONOUS_API)
|
|
{
|
|
/* Yes, this IRP was synchronous, so return the I/O Status */
|
|
*Irp->UserIosb = Irp->IoStatus;
|
|
|
|
/* Now check if the user gave an event */
|
|
if (Irp->UserEvent)
|
|
{
|
|
/* Signal it */
|
|
KeSetEvent(Irp->UserEvent, 0, FALSE);
|
|
}
|
|
else
|
|
{
|
|
/* No event was given, so signal the FO instead */
|
|
KeSetEvent(&FileObject->Event, 0, FALSE);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* It's not the IRP that was synchronous, it was the FO
|
|
* that was opened this way. Signal its event.
|
|
*/
|
|
FileObject->FinalStatus = Irp->IoStatus.Status;
|
|
KeSetEvent(&FileObject->Event, 0, FALSE);
|
|
}
|
|
}
|
|
|
|
/* Now that we got here, we do this for incomplete I/Os as well */
|
|
if ((FileObject) && !(Irp->Flags & IRP_CREATE_OPERATION))
|
|
{
|
|
/* Dereference the File Object unless this was a create */
|
|
ObDereferenceObjectDeferDelete(FileObject);
|
|
}
|
|
|
|
/*
|
|
* Check if this was an Executive Event (remember that we know this
|
|
* by checking if the IRP is synchronous)
|
|
*/
|
|
if ((Irp->UserEvent) &&
|
|
(FileObject) &&
|
|
!(Irp->Flags & IRP_SYNCHRONOUS_API))
|
|
{
|
|
/* This isn't a PKEVENT, so dereference it */
|
|
ObDereferenceObject(Irp->UserEvent);
|
|
}
|
|
|
|
/* Now that we've signaled the events, de-associate the IRP */
|
|
IopUnQueueIrpFromThread(Irp);
|
|
|
|
/* Free the IRP as well */
|
|
IoFreeIrp(Irp);
|
|
}
|
|
}
|
|
|
|
BOOLEAN
|
|
NTAPI
|
|
IopInitializeReserveIrp(IN PRESERVE_IRP_ALLOCATOR ReserveIrpAllocator)
|
|
{
|
|
/* Our allocated stack size */
|
|
ReserveIrpAllocator->StackSize = 20;
|
|
|
|
/* Allocate the IRP now */
|
|
ReserveIrpAllocator->ReserveIrp = IoAllocateIrp(ReserveIrpAllocator->StackSize, FALSE);
|
|
/* If we cannot, abort system boot */
|
|
if (ReserveIrpAllocator->ReserveIrp == NULL)
|
|
{
|
|
return FALSE;
|
|
}
|
|
|
|
/* It's not in use */
|
|
ReserveIrpAllocator->ReserveIrpInUse = 0;
|
|
/* And init the event */
|
|
KeInitializeEvent(&ReserveIrpAllocator->WaitEvent, SynchronizationEvent, FALSE);
|
|
|
|
/* All good, keep booting */
|
|
return TRUE;
|
|
}
|
|
|
|
PIRP
|
|
NTAPI
|
|
IopAllocateReserveIrp(IN CCHAR StackSize)
|
|
{
|
|
/* If we need a stack size higher than what was allocated, then fail */
|
|
if (StackSize > IopReserveIrpAllocator.StackSize)
|
|
{
|
|
return NULL;
|
|
}
|
|
|
|
/* Now, wait until the IRP becomes available and reserve it immediately */
|
|
while (InterlockedExchange(&IopReserveIrpAllocator.ReserveIrpInUse, 1) == 1)
|
|
{
|
|
KeWaitForSingleObject(&IopReserveIrpAllocator.WaitEvent,
|
|
Executive,
|
|
KernelMode,
|
|
FALSE,
|
|
NULL);
|
|
}
|
|
|
|
/* It's ours! Initialize it */
|
|
IoInitializeIrp(IopReserveIrpAllocator.ReserveIrp, IoSizeOfIrp(StackSize), StackSize);
|
|
|
|
/* And return it to the caller */
|
|
return IopReserveIrpAllocator.ReserveIrp;
|
|
}
|
|
|
|
VOID
|
|
IopFreeReserveIrp(IN CCHAR PriorityBoost)
|
|
{
|
|
/* Mark we don't use the IRP anymore */
|
|
InterlockedExchange(&IopReserveIrpAllocator.ReserveIrpInUse, 0);
|
|
|
|
/* And set the event if someone is waiting on the IRP */
|
|
KeSetEvent(&IopReserveIrpAllocator.WaitEvent, PriorityBoost, FALSE);
|
|
}
|
|
|
|
/* FUNCTIONS *****************************************************************/
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoAllocateIrp(IN CCHAR StackSize,
|
|
IN BOOLEAN ChargeQuota)
|
|
{
|
|
PIRP Irp = NULL;
|
|
USHORT Size = IoSizeOfIrp(StackSize);
|
|
PKPRCB Prcb;
|
|
UCHAR Flags = 0;
|
|
PNPAGED_LOOKASIDE_LIST List = NULL;
|
|
PP_NPAGED_LOOKASIDE_NUMBER ListType = LookasideSmallIrpList;
|
|
|
|
/* Set Charge Quota Flag */
|
|
if (ChargeQuota) Flags |= IRP_QUOTA_CHARGED;
|
|
|
|
/* Get the PRCB */
|
|
Prcb = KeGetCurrentPrcb();
|
|
|
|
/* Figure out which Lookaside List to use */
|
|
if ((StackSize <= 8) && (ChargeQuota == FALSE || Prcb->LookasideIrpFloat > 0))
|
|
{
|
|
/* Set Fixed Size Flag */
|
|
Flags |= IRP_ALLOCATED_FIXED_SIZE;
|
|
|
|
/* See if we should use big list */
|
|
if (StackSize != 1)
|
|
{
|
|
Size = IoSizeOfIrp(8);
|
|
ListType = LookasideLargeIrpList;
|
|
}
|
|
|
|
/* Get the P List First */
|
|
List = (PNPAGED_LOOKASIDE_LIST)Prcb->PPLookasideList[ListType].P;
|
|
|
|
/* Attempt allocation */
|
|
List->L.TotalAllocates++;
|
|
Irp = (PIRP)InterlockedPopEntrySList(&List->L.ListHead);
|
|
|
|
/* Check if the P List failed */
|
|
if (!Irp)
|
|
{
|
|
/* Let the balancer know */
|
|
List->L.AllocateMisses++;
|
|
|
|
/* Try the L List */
|
|
List = (PNPAGED_LOOKASIDE_LIST)Prcb->PPLookasideList[ListType].L;
|
|
List->L.TotalAllocates++;
|
|
Irp = (PIRP)InterlockedPopEntrySList(&List->L.ListHead);
|
|
}
|
|
}
|
|
|
|
/* Check if we have to use the pool */
|
|
if (!Irp)
|
|
{
|
|
/* Did we try lookaside and fail? */
|
|
if (Flags & IRP_ALLOCATED_FIXED_SIZE) List->L.AllocateMisses++;
|
|
|
|
/* Check if we should charge quota */
|
|
if (ChargeQuota)
|
|
{
|
|
Irp = ExAllocatePoolWithQuotaTag(NonPagedPool | POOL_QUOTA_FAIL_INSTEAD_OF_RAISE,
|
|
Size,
|
|
TAG_IRP);
|
|
}
|
|
else
|
|
{
|
|
/* Allocate the IRP with no quota charge */
|
|
Irp = ExAllocatePoolWithTag(NonPagedPool, Size, TAG_IRP);
|
|
}
|
|
|
|
/* Make sure it was sucessful */
|
|
if (!Irp) return NULL;
|
|
}
|
|
else if (Flags & IRP_QUOTA_CHARGED)
|
|
{
|
|
/* Decrement lookaside float */
|
|
InterlockedDecrement(&Prcb->LookasideIrpFloat);
|
|
Flags |= IRP_LOOKASIDE_ALLOCATION;
|
|
|
|
/* In this case there is no charge quota */
|
|
Flags &= ~IRP_QUOTA_CHARGED;
|
|
}
|
|
|
|
/* Now Initialize it */
|
|
IoInitializeIrp(Irp, Size, StackSize);
|
|
|
|
/* Set the Allocation Flags */
|
|
Irp->AllocationFlags = Flags;
|
|
|
|
/* Return it */
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Allocated IRP %p with allocation flags %lx\n",
|
|
__FUNCTION__,
|
|
Irp,
|
|
Flags);
|
|
return Irp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IopAllocateIrpMustSucceed(IN CCHAR StackSize)
|
|
{
|
|
LONG i;
|
|
PIRP Irp;
|
|
LARGE_INTEGER Sleep;
|
|
|
|
/* Try to get an IRP */
|
|
Irp = IoAllocateIrp(StackSize, FALSE);
|
|
if (Irp)
|
|
return Irp;
|
|
|
|
/* If we fail, start looping till we may get one */
|
|
i = LONG_MAX;
|
|
do {
|
|
i--;
|
|
|
|
/* First, sleep for 10ms */
|
|
Sleep.QuadPart = -10 * 1000 * 10;
|
|
KeDelayExecutionThread(KernelMode, FALSE, &Sleep);
|
|
|
|
/* Then, retry allocation */
|
|
Irp = IoAllocateIrp(StackSize, FALSE);
|
|
if (Irp)
|
|
return Irp;
|
|
} while (i > 0);
|
|
|
|
return Irp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoBuildAsynchronousFsdRequest(IN ULONG MajorFunction,
|
|
IN PDEVICE_OBJECT DeviceObject,
|
|
IN PVOID Buffer,
|
|
IN ULONG Length,
|
|
IN PLARGE_INTEGER StartingOffset,
|
|
IN PIO_STATUS_BLOCK IoStatusBlock)
|
|
{
|
|
PIRP Irp;
|
|
PIO_STACK_LOCATION StackPtr;
|
|
|
|
/* Allocate IRP */
|
|
Irp = IoAllocateIrp(DeviceObject->StackSize, FALSE);
|
|
if (!Irp) return NULL;
|
|
|
|
/* Get the Stack */
|
|
StackPtr = IoGetNextIrpStackLocation(Irp);
|
|
|
|
/* Write the Major function and then deal with it */
|
|
StackPtr->MajorFunction = (UCHAR)MajorFunction;
|
|
|
|
/* Do not handle the following here */
|
|
if ((MajorFunction != IRP_MJ_FLUSH_BUFFERS) &&
|
|
(MajorFunction != IRP_MJ_SHUTDOWN) &&
|
|
(MajorFunction != IRP_MJ_PNP) &&
|
|
(MajorFunction != IRP_MJ_POWER))
|
|
{
|
|
/* Check if this is Buffered IO */
|
|
if (DeviceObject->Flags & DO_BUFFERED_IO)
|
|
{
|
|
/* Allocate the System Buffer */
|
|
Irp->AssociatedIrp.SystemBuffer =
|
|
ExAllocatePoolWithTag(NonPagedPool, Length, TAG_SYS_BUF);
|
|
if (!Irp->AssociatedIrp.SystemBuffer)
|
|
{
|
|
/* Free the IRP and fail */
|
|
IoFreeIrp(Irp);
|
|
return NULL;
|
|
}
|
|
|
|
/* Set flags */
|
|
Irp->Flags = IRP_BUFFERED_IO | IRP_DEALLOCATE_BUFFER;
|
|
|
|
/* Handle special IRP_MJ_WRITE Case */
|
|
if (MajorFunction == IRP_MJ_WRITE)
|
|
{
|
|
/* Copy the buffer data */
|
|
RtlCopyMemory(Irp->AssociatedIrp.SystemBuffer, Buffer, Length);
|
|
}
|
|
else
|
|
{
|
|
/* Set the Input Operation flag and set this as a User Buffer */
|
|
Irp->Flags |= IRP_INPUT_OPERATION;
|
|
Irp->UserBuffer = Buffer;
|
|
}
|
|
}
|
|
else if (DeviceObject->Flags & DO_DIRECT_IO)
|
|
{
|
|
/* Use an MDL for Direct I/O */
|
|
Irp->MdlAddress = IoAllocateMdl(Buffer,
|
|
Length,
|
|
FALSE,
|
|
FALSE,
|
|
NULL);
|
|
if (!Irp->MdlAddress)
|
|
{
|
|
/* Free the IRP and fail */
|
|
IoFreeIrp(Irp);
|
|
return NULL;
|
|
}
|
|
|
|
/* Probe and Lock */
|
|
_SEH2_TRY
|
|
{
|
|
/* Do the probe */
|
|
MmProbeAndLockPages(Irp->MdlAddress,
|
|
KernelMode,
|
|
MajorFunction == IRP_MJ_READ ?
|
|
IoWriteAccess : IoReadAccess);
|
|
}
|
|
_SEH2_EXCEPT(EXCEPTION_EXECUTE_HANDLER)
|
|
{
|
|
/* Free the IRP and its MDL */
|
|
IoFreeMdl(Irp->MdlAddress);
|
|
IoFreeIrp(Irp);
|
|
|
|
/* Fail */
|
|
_SEH2_YIELD(return NULL);
|
|
}
|
|
_SEH2_END;
|
|
}
|
|
else
|
|
{
|
|
/* Neither, use the buffer */
|
|
Irp->UserBuffer = Buffer;
|
|
}
|
|
|
|
/* Check if this is a read */
|
|
if (MajorFunction == IRP_MJ_READ)
|
|
{
|
|
/* Set the parameters for a read */
|
|
StackPtr->Parameters.Read.Length = Length;
|
|
StackPtr->Parameters.Read.ByteOffset = *StartingOffset;
|
|
}
|
|
else if (MajorFunction == IRP_MJ_WRITE)
|
|
{
|
|
/* Otherwise, set write parameters */
|
|
StackPtr->Parameters.Write.Length = Length;
|
|
StackPtr->Parameters.Write.ByteOffset = *StartingOffset;
|
|
}
|
|
}
|
|
|
|
/* Set the Current Thread and IOSB */
|
|
Irp->UserIosb = IoStatusBlock;
|
|
Irp->Tail.Overlay.Thread = PsGetCurrentThread();
|
|
|
|
/* Return the IRP */
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Built IRP %p with Major, Buffer, DO %lx %p %p\n",
|
|
__FUNCTION__,
|
|
Irp,
|
|
MajorFunction,
|
|
Buffer,
|
|
DeviceObject);
|
|
return Irp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoBuildDeviceIoControlRequest(IN ULONG IoControlCode,
|
|
IN PDEVICE_OBJECT DeviceObject,
|
|
IN PVOID InputBuffer,
|
|
IN ULONG InputBufferLength,
|
|
IN PVOID OutputBuffer,
|
|
IN ULONG OutputBufferLength,
|
|
IN BOOLEAN InternalDeviceIoControl,
|
|
IN PKEVENT Event,
|
|
IN PIO_STATUS_BLOCK IoStatusBlock)
|
|
{
|
|
PIRP Irp;
|
|
PIO_STACK_LOCATION StackPtr;
|
|
ULONG BufferLength;
|
|
|
|
/* Allocate IRP */
|
|
Irp = IoAllocateIrp(DeviceObject->StackSize, FALSE);
|
|
if (!Irp) return NULL;
|
|
|
|
/* Get the Stack */
|
|
StackPtr = IoGetNextIrpStackLocation(Irp);
|
|
|
|
/* Set the DevCtl Type */
|
|
StackPtr->MajorFunction = InternalDeviceIoControl ?
|
|
IRP_MJ_INTERNAL_DEVICE_CONTROL :
|
|
IRP_MJ_DEVICE_CONTROL;
|
|
|
|
/* Set the IOCTL Data */
|
|
StackPtr->Parameters.DeviceIoControl.IoControlCode = IoControlCode;
|
|
StackPtr->Parameters.DeviceIoControl.InputBufferLength = InputBufferLength;
|
|
StackPtr->Parameters.DeviceIoControl.OutputBufferLength =
|
|
OutputBufferLength;
|
|
|
|
/* Handle the Methods */
|
|
switch (IO_METHOD_FROM_CTL_CODE(IoControlCode))
|
|
{
|
|
/* Buffered I/O */
|
|
case METHOD_BUFFERED:
|
|
|
|
/* Select the right Buffer Length */
|
|
BufferLength = InputBufferLength > OutputBufferLength ?
|
|
InputBufferLength : OutputBufferLength;
|
|
|
|
/* Make sure there is one */
|
|
if (BufferLength)
|
|
{
|
|
/* Allocate the System Buffer */
|
|
Irp->AssociatedIrp.SystemBuffer =
|
|
ExAllocatePoolWithTag(NonPagedPool,
|
|
BufferLength,
|
|
TAG_SYS_BUF);
|
|
if (!Irp->AssociatedIrp.SystemBuffer)
|
|
{
|
|
/* Free the IRP and fail */
|
|
IoFreeIrp(Irp);
|
|
return NULL;
|
|
}
|
|
|
|
/* Check if we got a buffer */
|
|
if (InputBuffer)
|
|
{
|
|
/* Copy into the System Buffer */
|
|
RtlCopyMemory(Irp->AssociatedIrp.SystemBuffer,
|
|
InputBuffer,
|
|
InputBufferLength);
|
|
}
|
|
|
|
/* Write the flags */
|
|
Irp->Flags = IRP_BUFFERED_IO | IRP_DEALLOCATE_BUFFER;
|
|
if (OutputBuffer) Irp->Flags |= IRP_INPUT_OPERATION;
|
|
|
|
/* Save the Buffer */
|
|
Irp->UserBuffer = OutputBuffer;
|
|
}
|
|
else
|
|
{
|
|
/* Clear the Flags and Buffer */
|
|
Irp->Flags = 0;
|
|
Irp->UserBuffer = NULL;
|
|
}
|
|
break;
|
|
|
|
/* Direct I/O */
|
|
case METHOD_IN_DIRECT:
|
|
case METHOD_OUT_DIRECT:
|
|
|
|
/* Check if we got an input buffer */
|
|
if (InputBuffer)
|
|
{
|
|
/* Allocate the System Buffer */
|
|
Irp->AssociatedIrp.SystemBuffer =
|
|
ExAllocatePoolWithTag(NonPagedPool,
|
|
InputBufferLength,
|
|
TAG_SYS_BUF);
|
|
if (!Irp->AssociatedIrp.SystemBuffer)
|
|
{
|
|
/* Free the IRP and fail */
|
|
IoFreeIrp(Irp);
|
|
return NULL;
|
|
}
|
|
|
|
/* Copy into the System Buffer */
|
|
RtlCopyMemory(Irp->AssociatedIrp.SystemBuffer,
|
|
InputBuffer,
|
|
InputBufferLength);
|
|
|
|
/* Write the flags */
|
|
Irp->Flags = IRP_BUFFERED_IO | IRP_DEALLOCATE_BUFFER;
|
|
}
|
|
else
|
|
{
|
|
/* Clear the flags */
|
|
Irp->Flags = 0;
|
|
}
|
|
|
|
/* Check if we got an output buffer */
|
|
if (OutputBuffer)
|
|
{
|
|
/* Allocate the System Buffer */
|
|
Irp->MdlAddress = IoAllocateMdl(OutputBuffer,
|
|
OutputBufferLength,
|
|
FALSE,
|
|
FALSE,
|
|
Irp);
|
|
if (!Irp->MdlAddress)
|
|
{
|
|
/* Free the IRP and fail */
|
|
IoFreeIrp(Irp);
|
|
return NULL;
|
|
}
|
|
|
|
/* Probe and Lock */
|
|
_SEH2_TRY
|
|
{
|
|
/* Do the probe */
|
|
MmProbeAndLockPages(Irp->MdlAddress,
|
|
KernelMode,
|
|
IO_METHOD_FROM_CTL_CODE(IoControlCode) ==
|
|
METHOD_IN_DIRECT ?
|
|
IoReadAccess : IoWriteAccess);
|
|
}
|
|
_SEH2_EXCEPT(EXCEPTION_EXECUTE_HANDLER)
|
|
{
|
|
/* Free the MDL */
|
|
IoFreeMdl(Irp->MdlAddress);
|
|
|
|
/* Free the input buffer and IRP */
|
|
if (InputBuffer) ExFreePool(Irp->AssociatedIrp.SystemBuffer);
|
|
IoFreeIrp(Irp);
|
|
|
|
/* Fail */
|
|
_SEH2_YIELD(return NULL);
|
|
}
|
|
_SEH2_END;
|
|
}
|
|
break;
|
|
|
|
case METHOD_NEITHER:
|
|
|
|
/* Just save the Buffer */
|
|
Irp->UserBuffer = OutputBuffer;
|
|
StackPtr->Parameters.DeviceIoControl.Type3InputBuffer = InputBuffer;
|
|
}
|
|
|
|
/* Now write the Event and IoSB */
|
|
Irp->UserIosb = IoStatusBlock;
|
|
Irp->UserEvent = Event;
|
|
|
|
/* Sync IRPs are queued to requestor thread's irp cancel/cleanup list */
|
|
Irp->Tail.Overlay.Thread = PsGetCurrentThread();
|
|
IoQueueThreadIrp(Irp);
|
|
|
|
/* Return the IRP */
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Built IRP %p with IOCTL, Buffers, DO %lx %p %p %p\n",
|
|
__FUNCTION__,
|
|
Irp,
|
|
IoControlCode,
|
|
InputBuffer,
|
|
OutputBuffer,
|
|
DeviceObject);
|
|
return Irp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoBuildSynchronousFsdRequest(IN ULONG MajorFunction,
|
|
IN PDEVICE_OBJECT DeviceObject,
|
|
IN PVOID Buffer,
|
|
IN ULONG Length,
|
|
IN PLARGE_INTEGER StartingOffset,
|
|
IN PKEVENT Event,
|
|
IN PIO_STATUS_BLOCK IoStatusBlock)
|
|
{
|
|
PIRP Irp;
|
|
|
|
/* Do the big work to set up the IRP */
|
|
Irp = IoBuildAsynchronousFsdRequest(MajorFunction,
|
|
DeviceObject,
|
|
Buffer,
|
|
Length,
|
|
StartingOffset,
|
|
IoStatusBlock );
|
|
if (!Irp) return NULL;
|
|
|
|
/* Set the Event which makes it Syncronous */
|
|
Irp->UserEvent = Event;
|
|
|
|
/* Sync IRPs are queued to requestor thread's irp cancel/cleanup list */
|
|
IoQueueThreadIrp(Irp);
|
|
return Irp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
BOOLEAN
|
|
NTAPI
|
|
IoCancelIrp(IN PIRP Irp)
|
|
{
|
|
KIRQL OldIrql;
|
|
PDRIVER_CANCEL CancelRoutine;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Canceling IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
ASSERT(Irp->Type == IO_TYPE_IRP);
|
|
|
|
/* Acquire the cancel lock and cancel the IRP */
|
|
IoAcquireCancelSpinLock(&OldIrql);
|
|
Irp->Cancel = TRUE;
|
|
|
|
/* Clear the cancel routine and get the old one */
|
|
CancelRoutine = IoSetCancelRoutine(Irp, NULL);
|
|
if (CancelRoutine)
|
|
{
|
|
/* We had a routine, make sure the IRP isn't completed */
|
|
if (Irp->CurrentLocation > (Irp->StackCount + 1))
|
|
{
|
|
/* It is, bugcheck */
|
|
KeBugCheckEx(CANCEL_STATE_IN_COMPLETED_IRP,
|
|
(ULONG_PTR)Irp,
|
|
(ULONG_PTR)CancelRoutine,
|
|
0,
|
|
0);
|
|
}
|
|
|
|
/* Set the cancel IRQL And call the routine */
|
|
Irp->CancelIrql = OldIrql;
|
|
CancelRoutine(IoGetCurrentIrpStackLocation(Irp)->DeviceObject, Irp);
|
|
return TRUE;
|
|
}
|
|
|
|
/* Otherwise, release the cancel lock and fail */
|
|
IoReleaseCancelSpinLock(OldIrql);
|
|
return FALSE;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoCancelThreadIo(IN PETHREAD Thread)
|
|
{
|
|
KIRQL OldIrql;
|
|
ULONG Retries = 3000;
|
|
LARGE_INTEGER Interval;
|
|
PLIST_ENTRY ListHead, NextEntry;
|
|
PIRP Irp;
|
|
PAGED_CODE();
|
|
|
|
/* Windows isn't using given thread, but using current. */
|
|
Thread = PsGetCurrentThread();
|
|
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Canceling IRPs for Thread %p\n",
|
|
__FUNCTION__,
|
|
Thread);
|
|
|
|
/* Raise to APC to protect the IrpList */
|
|
KeRaiseIrql(APC_LEVEL, &OldIrql);
|
|
|
|
/* Start by cancelling all the IRPs in the current thread queue. */
|
|
ListHead = &Thread->IrpList;
|
|
NextEntry = ListHead->Flink;
|
|
while (ListHead != NextEntry)
|
|
{
|
|
/* Get the IRP */
|
|
Irp = CONTAINING_RECORD(NextEntry, IRP, ThreadListEntry);
|
|
|
|
/* Cancel it */
|
|
IoCancelIrp(Irp);
|
|
|
|
/* Move to the next entry */
|
|
NextEntry = NextEntry->Flink;
|
|
}
|
|
|
|
/* Wait 100 milliseconds */
|
|
Interval.QuadPart = -1000000;
|
|
|
|
/* Wait till all the IRPs are completed or cancelled. */
|
|
while (!IsListEmpty(&Thread->IrpList))
|
|
{
|
|
/* Now we can lower */
|
|
KeLowerIrql(OldIrql);
|
|
|
|
/* Wait a short while and then look if all our IRPs were completed. */
|
|
KeDelayExecutionThread(KernelMode, FALSE, &Interval);
|
|
|
|
/*
|
|
* Don't stay here forever if some broken driver doesn't complete
|
|
* the IRP.
|
|
*/
|
|
if (!(Retries--))
|
|
{
|
|
/* Print out a message and remove the IRP */
|
|
DPRINT1("Broken driver did not complete!\n");
|
|
IopDisassociateThreadIrp();
|
|
}
|
|
|
|
/* Raise the IRQL Again */
|
|
KeRaiseIrql(APC_LEVEL, &OldIrql);
|
|
}
|
|
|
|
/* We're done, lower the IRQL */
|
|
KeLowerIrql(OldIrql);
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
#undef IoCallDriver
|
|
NTSTATUS
|
|
NTAPI
|
|
IoCallDriver(IN PDEVICE_OBJECT DeviceObject,
|
|
IN PIRP Irp)
|
|
{
|
|
/* Call fastcall */
|
|
return IofCallDriver(DeviceObject, Irp);
|
|
}
|
|
|
|
#define IoCallDriver IofCallDriver
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
#undef IoCompleteRequest
|
|
VOID
|
|
NTAPI
|
|
IoCompleteRequest(IN PIRP Irp,
|
|
IN CCHAR PriorityBoost)
|
|
{
|
|
/* Call the fastcall */
|
|
IofCompleteRequest(Irp, PriorityBoost);
|
|
}
|
|
|
|
#define IoCompleteRequest IofCompleteRequest
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoEnqueueIrp(IN PIRP Irp)
|
|
{
|
|
/* This is the same as calling IoQueueThreadIrp */
|
|
IoQueueThreadIrp(Irp);
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
NTSTATUS
|
|
FASTCALL
|
|
IofCallDriver(IN PDEVICE_OBJECT DeviceObject,
|
|
IN PIRP Irp)
|
|
{
|
|
PDRIVER_OBJECT DriverObject;
|
|
PIO_STACK_LOCATION StackPtr;
|
|
|
|
/* Make sure this is a valid IRP */
|
|
ASSERT(Irp->Type == IO_TYPE_IRP);
|
|
|
|
/* Get the Driver Object */
|
|
DriverObject = DeviceObject->DriverObject;
|
|
|
|
/* Decrease the current location and check if */
|
|
Irp->CurrentLocation--;
|
|
if (Irp->CurrentLocation <= 0)
|
|
{
|
|
/* This IRP ran out of stack, bugcheck */
|
|
KeBugCheckEx(NO_MORE_IRP_STACK_LOCATIONS, (ULONG_PTR)Irp, 0, 0, 0);
|
|
}
|
|
|
|
/* Now update the stack location */
|
|
StackPtr = IoGetNextIrpStackLocation(Irp);
|
|
Irp->Tail.Overlay.CurrentStackLocation = StackPtr;
|
|
|
|
/* Get the Device Object */
|
|
StackPtr->DeviceObject = DeviceObject;
|
|
|
|
/* Call it */
|
|
return DriverObject->MajorFunction[StackPtr->MajorFunction](DeviceObject,
|
|
Irp);
|
|
}
|
|
|
|
FORCEINLINE
|
|
VOID
|
|
IopClearStackLocation(IN PIO_STACK_LOCATION IoStackLocation)
|
|
{
|
|
IoStackLocation->MinorFunction = 0;
|
|
IoStackLocation->Flags = 0;
|
|
IoStackLocation->Control &= SL_ERROR_RETURNED;
|
|
IoStackLocation->Parameters.Others.Argument1 = 0;
|
|
IoStackLocation->Parameters.Others.Argument2 = 0;
|
|
IoStackLocation->Parameters.Others.Argument3 = 0;
|
|
IoStackLocation->FileObject = NULL;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
FASTCALL
|
|
IofCompleteRequest(IN PIRP Irp,
|
|
IN CCHAR PriorityBoost)
|
|
{
|
|
PIO_STACK_LOCATION StackPtr, LastStackPtr;
|
|
PDEVICE_OBJECT DeviceObject;
|
|
PFILE_OBJECT FileObject;
|
|
PETHREAD Thread;
|
|
NTSTATUS Status;
|
|
PMDL Mdl, NextMdl;
|
|
ULONG MasterCount;
|
|
PIRP MasterIrp;
|
|
ULONG Flags;
|
|
NTSTATUS ErrorCode = STATUS_SUCCESS;
|
|
PREPARSE_DATA_BUFFER DataBuffer = NULL;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Completing IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
|
|
/* Make sure this IRP isn't getting completed twice or is invalid */
|
|
if ((Irp->CurrentLocation) > (Irp->StackCount + 1))
|
|
{
|
|
/* Bugcheck */
|
|
KeBugCheckEx(MULTIPLE_IRP_COMPLETE_REQUESTS, (ULONG_PTR)Irp, 0, 0, 0);
|
|
}
|
|
|
|
/* Some sanity checks */
|
|
ASSERT(Irp->Type == IO_TYPE_IRP);
|
|
ASSERT(!Irp->CancelRoutine);
|
|
ASSERT(Irp->IoStatus.Status != STATUS_PENDING);
|
|
ASSERT(Irp->IoStatus.Status != (NTSTATUS)0xFFFFFFFF);
|
|
|
|
/* Get the last stack */
|
|
LastStackPtr = (PIO_STACK_LOCATION)(Irp + 1);
|
|
if (LastStackPtr->Control & SL_ERROR_RETURNED)
|
|
{
|
|
/* Get the error code */
|
|
ErrorCode = PtrToUlong(LastStackPtr->Parameters.Others.Argument4);
|
|
}
|
|
|
|
/*
|
|
* Start the loop with the current stack and point the IRP to the next stack
|
|
* and then keep incrementing the stack as we loop through. The IRP should
|
|
* always point to the next stack location w.r.t the one currently being
|
|
* analyzed, so completion routine code will see the appropriate value.
|
|
* Because of this, we must loop until the current stack location is +1 of
|
|
* the stack count, because when StackPtr is at the end, CurrentLocation is +1.
|
|
*/
|
|
for (StackPtr = IoGetCurrentIrpStackLocation(Irp),
|
|
Irp->CurrentLocation++,
|
|
Irp->Tail.Overlay.CurrentStackLocation++;
|
|
Irp->CurrentLocation <= (Irp->StackCount + 1);
|
|
StackPtr++,
|
|
Irp->CurrentLocation++,
|
|
Irp->Tail.Overlay.CurrentStackLocation++)
|
|
{
|
|
/* Set Pending Returned */
|
|
Irp->PendingReturned = StackPtr->Control & SL_PENDING_RETURNED;
|
|
|
|
/* Check if we failed */
|
|
if (!NT_SUCCESS(Irp->IoStatus.Status))
|
|
{
|
|
/* Check if it was changed by a completion routine */
|
|
if (Irp->IoStatus.Status != ErrorCode)
|
|
{
|
|
/* Update the error for the current stack */
|
|
ErrorCode = Irp->IoStatus.Status;
|
|
StackPtr->Control |= SL_ERROR_RETURNED;
|
|
LastStackPtr->Parameters.Others.Argument4 = UlongToPtr(ErrorCode);
|
|
LastStackPtr->Control |= SL_ERROR_RETURNED;
|
|
}
|
|
}
|
|
|
|
/* Check if there is a Completion Routine to Call */
|
|
if ((NT_SUCCESS(Irp->IoStatus.Status) &&
|
|
(StackPtr->Control & SL_INVOKE_ON_SUCCESS)) ||
|
|
(!NT_SUCCESS(Irp->IoStatus.Status) &&
|
|
(StackPtr->Control & SL_INVOKE_ON_ERROR)) ||
|
|
(Irp->Cancel &&
|
|
(StackPtr->Control & SL_INVOKE_ON_CANCEL)))
|
|
{
|
|
/* Clear the stack location */
|
|
IopClearStackLocation(StackPtr);
|
|
|
|
/* Check for highest-level device completion routines */
|
|
if (Irp->CurrentLocation == (Irp->StackCount + 1))
|
|
{
|
|
/* Clear the DO, since the current stack location is invalid */
|
|
DeviceObject = NULL;
|
|
}
|
|
else
|
|
{
|
|
/* Otherwise, return the real one */
|
|
DeviceObject = IoGetCurrentIrpStackLocation(Irp)->DeviceObject;
|
|
}
|
|
|
|
/* Call the completion routine */
|
|
Status = StackPtr->CompletionRoutine(DeviceObject,
|
|
Irp,
|
|
StackPtr->Context);
|
|
|
|
/* Don't touch the Packet in this case, since it might be gone! */
|
|
if (Status == STATUS_MORE_PROCESSING_REQUIRED) return;
|
|
}
|
|
else
|
|
{
|
|
/* Otherwise, check if this is a completed IRP */
|
|
if ((Irp->CurrentLocation <= Irp->StackCount) &&
|
|
(Irp->PendingReturned))
|
|
{
|
|
/* Mark it as pending */
|
|
IoMarkIrpPending(Irp);
|
|
}
|
|
|
|
/* Clear the stack location */
|
|
IopClearStackLocation(StackPtr);
|
|
}
|
|
}
|
|
|
|
/* Check if the IRP is an associated IRP */
|
|
if (Irp->Flags & IRP_ASSOCIATED_IRP)
|
|
{
|
|
/* Get the master IRP and count */
|
|
MasterIrp = Irp->AssociatedIrp.MasterIrp;
|
|
MasterCount = InterlockedDecrement(&MasterIrp->AssociatedIrp.IrpCount);
|
|
|
|
/* Free the MDLs */
|
|
for (Mdl = Irp->MdlAddress; Mdl; Mdl = NextMdl)
|
|
{
|
|
/* Go to the next one */
|
|
NextMdl = Mdl->Next;
|
|
IoFreeMdl(Mdl);
|
|
}
|
|
|
|
/* Free the IRP itself */
|
|
IoFreeIrp(Irp);
|
|
|
|
/* Complete the Master IRP */
|
|
if (!MasterCount) IofCompleteRequest(MasterIrp, PriorityBoost);
|
|
return;
|
|
}
|
|
|
|
/* Check whether we have to reparse */
|
|
if (Irp->IoStatus.Status == STATUS_REPARSE)
|
|
{
|
|
if (Irp->IoStatus.Information > IO_REMOUNT)
|
|
{
|
|
/* If that's a reparse tag we understand, save the buffer from deletion */
|
|
if (Irp->IoStatus.Information == IO_REPARSE_TAG_MOUNT_POINT)
|
|
{
|
|
ASSERT(Irp->Tail.Overlay.AuxiliaryBuffer != NULL);
|
|
DataBuffer = (PREPARSE_DATA_BUFFER)Irp->Tail.Overlay.AuxiliaryBuffer;
|
|
Irp->Tail.Overlay.AuxiliaryBuffer = NULL;
|
|
}
|
|
else
|
|
{
|
|
Irp->IoStatus.Status = STATUS_IO_REPARSE_TAG_NOT_HANDLED;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Check if we have an auxiliary buffer */
|
|
if (Irp->Tail.Overlay.AuxiliaryBuffer)
|
|
{
|
|
/* Free it */
|
|
ExFreePool(Irp->Tail.Overlay.AuxiliaryBuffer);
|
|
Irp->Tail.Overlay.AuxiliaryBuffer = NULL;
|
|
}
|
|
|
|
/* Check if this is a Paging I/O or Close Operation */
|
|
if (Irp->Flags & (IRP_PAGING_IO | IRP_CLOSE_OPERATION))
|
|
{
|
|
/* Handle a Close Operation or Sync Paging I/O */
|
|
if (Irp->Flags & (IRP_SYNCHRONOUS_PAGING_IO | IRP_CLOSE_OPERATION))
|
|
{
|
|
/* Set the I/O Status and Signal the Event */
|
|
Flags = Irp->Flags & (IRP_SYNCHRONOUS_PAGING_IO | IRP_PAGING_IO);
|
|
*Irp->UserIosb = Irp->IoStatus;
|
|
KeSetEvent(Irp->UserEvent, PriorityBoost, FALSE);
|
|
|
|
/* Free the IRP for a Paging I/O Only, Close is handled by us */
|
|
if (Flags)
|
|
{
|
|
/* If we were using the reserve IRP, then call the appropriate
|
|
* free function (to make the IRP available again)
|
|
*/
|
|
if (Irp == IopReserveIrpAllocator.ReserveIrp)
|
|
{
|
|
IopFreeReserveIrp(PriorityBoost);
|
|
}
|
|
/* Otherwise, free for real! */
|
|
else
|
|
{
|
|
IoFreeIrp(Irp);
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
#if 0
|
|
/* Page 166 */
|
|
KeInitializeApc(&Irp->Tail.Apc
|
|
&Irp->Tail.Overlay.Thread->Tcb,
|
|
Irp->ApcEnvironment,
|
|
IopCompletePageWrite,
|
|
NULL,
|
|
NULL,
|
|
KernelMode,
|
|
NULL);
|
|
KeInsertQueueApc(&Irp->Tail.Apc,
|
|
NULL,
|
|
NULL,
|
|
PriorityBoost);
|
|
#else
|
|
/* Not implemented yet. */
|
|
UNIMPLEMENTED_DBGBREAK("Not supported!\n");
|
|
#endif
|
|
}
|
|
|
|
/* Get out of here */
|
|
return;
|
|
}
|
|
|
|
/* Unlock MDL Pages, page 167. */
|
|
Mdl = Irp->MdlAddress;
|
|
while (Mdl)
|
|
{
|
|
MmUnlockPages(Mdl);
|
|
Mdl = Mdl->Next;
|
|
}
|
|
|
|
/* Check if we should exit because of a Deferred I/O (page 168) */
|
|
if ((Irp->Flags & IRP_DEFER_IO_COMPLETION) && !(Irp->PendingReturned))
|
|
{
|
|
/* Restore the saved reparse buffer for the caller */
|
|
if (Irp->IoStatus.Status == STATUS_REPARSE &&
|
|
Irp->IoStatus.Information == IO_REPARSE_TAG_MOUNT_POINT)
|
|
{
|
|
Irp->Tail.Overlay.AuxiliaryBuffer = (PCHAR)DataBuffer;
|
|
}
|
|
|
|
/*
|
|
* Return without queuing the completion APC, since the caller will
|
|
* take care of doing its own optimized completion at PASSIVE_LEVEL.
|
|
*/
|
|
return;
|
|
}
|
|
|
|
/* Get the thread and file object */
|
|
Thread = Irp->Tail.Overlay.Thread;
|
|
FileObject = Irp->Tail.Overlay.OriginalFileObject;
|
|
|
|
/* Make sure the IRP isn't canceled */
|
|
if (!Irp->Cancel)
|
|
{
|
|
/* Initialize the APC */
|
|
KeInitializeApc(&Irp->Tail.Apc,
|
|
&Thread->Tcb,
|
|
Irp->ApcEnvironment,
|
|
IopCompleteRequest,
|
|
NULL,
|
|
NULL,
|
|
KernelMode,
|
|
NULL);
|
|
|
|
/* Queue it */
|
|
KeInsertQueueApc(&Irp->Tail.Apc,
|
|
FileObject,
|
|
DataBuffer,
|
|
PriorityBoost);
|
|
}
|
|
else
|
|
{
|
|
/* The IRP just got canceled... does a thread still own it? */
|
|
if (Thread)
|
|
{
|
|
/* Yes! There is still hope! Initialize the APC */
|
|
KeInitializeApc(&Irp->Tail.Apc,
|
|
&Thread->Tcb,
|
|
Irp->ApcEnvironment,
|
|
IopCompleteRequest,
|
|
NULL,
|
|
NULL,
|
|
KernelMode,
|
|
NULL);
|
|
|
|
/* Queue it */
|
|
KeInsertQueueApc(&Irp->Tail.Apc,
|
|
FileObject,
|
|
DataBuffer,
|
|
PriorityBoost);
|
|
}
|
|
else
|
|
{
|
|
/* Nothing left for us to do, kill it */
|
|
ASSERT(Irp->Cancel);
|
|
IopCleanupIrp(Irp, FileObject);
|
|
}
|
|
}
|
|
}
|
|
|
|
NTSTATUS
|
|
NTAPI
|
|
IopSynchronousCompletion(IN PDEVICE_OBJECT DeviceObject,
|
|
IN PIRP Irp,
|
|
IN PVOID Context)
|
|
{
|
|
if (Irp->PendingReturned)
|
|
KeSetEvent((PKEVENT)Context, IO_NO_INCREMENT, FALSE);
|
|
return STATUS_MORE_PROCESSING_REQUIRED;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
BOOLEAN
|
|
NTAPI
|
|
IoForwardIrpSynchronously(IN PDEVICE_OBJECT DeviceObject,
|
|
IN PIRP Irp)
|
|
{
|
|
KEVENT Event;
|
|
NTSTATUS Status;
|
|
|
|
/* Check if next stack location is available */
|
|
if (Irp->CurrentLocation < Irp->StackCount)
|
|
{
|
|
/* No more stack location */
|
|
return FALSE;
|
|
}
|
|
|
|
/* Initialize event */
|
|
KeInitializeEvent(&Event, NotificationEvent, FALSE);
|
|
|
|
/* Copy stack location for next driver */
|
|
IoCopyCurrentIrpStackLocationToNext(Irp);
|
|
|
|
/* Set a completion routine, which will signal the event */
|
|
IoSetCompletionRoutine(Irp, IopSynchronousCompletion, &Event, TRUE, TRUE, TRUE);
|
|
|
|
/* Call next driver */
|
|
Status = IoCallDriver(DeviceObject, Irp);
|
|
|
|
/* Check if irp is pending */
|
|
if (Status == STATUS_PENDING)
|
|
{
|
|
/* Yes, wait for its completion */
|
|
KeWaitForSingleObject(&Event, Suspended, KernelMode, FALSE, NULL);
|
|
}
|
|
|
|
/* Return success */
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoFreeIrp(IN PIRP Irp)
|
|
{
|
|
PNPAGED_LOOKASIDE_LIST List;
|
|
PP_NPAGED_LOOKASIDE_NUMBER ListType = LookasideSmallIrpList;
|
|
PKPRCB Prcb;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Freeing IRPs %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
|
|
/* Make sure the Thread IRP list is empty and that it OK to free it */
|
|
ASSERT(Irp->Type == IO_TYPE_IRP);
|
|
ASSERT(IsListEmpty(&Irp->ThreadListEntry));
|
|
ASSERT(Irp->CurrentLocation >= Irp->StackCount);
|
|
|
|
/* Get the PRCB */
|
|
Prcb = KeGetCurrentPrcb();
|
|
|
|
/* If this was a lookaside alloc, increment lookaside float */
|
|
if (Irp->AllocationFlags & IRP_LOOKASIDE_ALLOCATION)
|
|
{
|
|
Irp->AllocationFlags &= ~IRP_LOOKASIDE_ALLOCATION;
|
|
InterlockedIncrement(&Prcb->LookasideIrpFloat);
|
|
}
|
|
|
|
/* If this was a pool alloc, free it with the pool */
|
|
if (!(Irp->AllocationFlags & IRP_ALLOCATED_FIXED_SIZE))
|
|
{
|
|
/* Free it */
|
|
ExFreePoolWithTag(Irp, TAG_IRP);
|
|
}
|
|
else
|
|
{
|
|
/* Check if this was a Big IRP */
|
|
if (Irp->StackCount != 1) ListType = LookasideLargeIrpList;
|
|
|
|
/* Use the P List */
|
|
List = (PNPAGED_LOOKASIDE_LIST)Prcb->PPLookasideList[ListType].P;
|
|
List->L.TotalFrees++;
|
|
|
|
/* Check if the Free was within the Depth or not */
|
|
if (ExQueryDepthSList(&List->L.ListHead) >= List->L.Depth)
|
|
{
|
|
/* Let the balancer know */
|
|
List->L.FreeMisses++;
|
|
|
|
/* Use the L List */
|
|
List = (PNPAGED_LOOKASIDE_LIST)Prcb->PPLookasideList[ListType].L;
|
|
List->L.TotalFrees++;
|
|
|
|
/* Check if the Free was within the Depth or not */
|
|
if (ExQueryDepthSList(&List->L.ListHead) >= List->L.Depth)
|
|
{
|
|
/* All lists failed, use the pool */
|
|
List->L.FreeMisses++;
|
|
ExFreePoolWithTag(Irp, TAG_IRP);
|
|
Irp = NULL;
|
|
}
|
|
}
|
|
|
|
/* The free was within the Depth */
|
|
if (Irp)
|
|
{
|
|
/* Remove the association with the process */
|
|
if (Irp->AllocationFlags & IRP_QUOTA_CHARGED)
|
|
{
|
|
ExReturnPoolQuota(Irp);
|
|
Irp->AllocationFlags &= ~IRP_QUOTA_CHARGED;
|
|
}
|
|
|
|
/* Add it to the lookaside list */
|
|
InterlockedPushEntrySList(&List->L.ListHead,
|
|
(PSLIST_ENTRY)Irp);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
IO_PAGING_PRIORITY
|
|
FASTCALL
|
|
IoGetPagingIoPriority(IN PIRP Irp)
|
|
{
|
|
IO_PAGING_PRIORITY Priority;
|
|
ULONG Flags;
|
|
|
|
/* Get the flags */
|
|
Flags = Irp->Flags;
|
|
|
|
/* Check what priority it has */
|
|
if (Flags & IRP_CLASS_CACHE_OPERATION)
|
|
{
|
|
/* High priority */
|
|
Priority = IoPagingPriorityHigh;
|
|
}
|
|
else if (Flags & IRP_PAGING_IO)
|
|
{
|
|
/* Normal priority */
|
|
Priority = IoPagingPriorityNormal;
|
|
}
|
|
else
|
|
{
|
|
/* Invalid -- not a paging IRP */
|
|
Priority = IoPagingPriorityInvalid;
|
|
}
|
|
|
|
/* Return the priority */
|
|
return Priority;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PEPROCESS
|
|
NTAPI
|
|
IoGetRequestorProcess(IN PIRP Irp)
|
|
{
|
|
/* Return the requestor process */
|
|
if (Irp->Tail.Overlay.Thread)
|
|
{
|
|
if (Irp->ApcEnvironment == OriginalApcEnvironment)
|
|
{
|
|
return Irp->Tail.Overlay.Thread->ThreadsProcess;
|
|
}
|
|
else if (Irp->ApcEnvironment == AttachedApcEnvironment)
|
|
{
|
|
return (PEPROCESS)Irp->Tail.Overlay.Thread->Tcb.ApcState.Process;
|
|
}
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
ULONG
|
|
NTAPI
|
|
IoGetRequestorProcessId(IN PIRP Irp)
|
|
{
|
|
PEPROCESS Process;
|
|
|
|
/* Return the requestor process' id */
|
|
Process = IoGetRequestorProcess(Irp);
|
|
if (Process) return PtrToUlong(Process->UniqueProcessId);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
NTSTATUS
|
|
NTAPI
|
|
IoGetRequestorSessionId(IN PIRP Irp,
|
|
OUT PULONG pSessionId)
|
|
{
|
|
PEPROCESS Process;
|
|
|
|
/* Return the session */
|
|
if (Irp->Tail.Overlay.Thread)
|
|
{
|
|
Process = Irp->Tail.Overlay.Thread->ThreadsProcess;
|
|
*pSessionId = MmGetSessionId(Process);
|
|
return STATUS_SUCCESS;
|
|
}
|
|
|
|
*pSessionId = (ULONG)-1;
|
|
return STATUS_UNSUCCESSFUL;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoGetTopLevelIrp(VOID)
|
|
{
|
|
/* Return the IRP */
|
|
return (PIRP)PsGetCurrentThread()->TopLevelIrp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoInitializeIrp(IN PIRP Irp,
|
|
IN USHORT PacketSize,
|
|
IN CCHAR StackSize)
|
|
{
|
|
/* Clear it */
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Initializing IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
RtlZeroMemory(Irp, PacketSize);
|
|
|
|
/* Set the Header and other data */
|
|
Irp->Type = IO_TYPE_IRP;
|
|
Irp->Size = PacketSize;
|
|
Irp->StackCount = StackSize;
|
|
Irp->CurrentLocation = StackSize + 1;
|
|
Irp->ApcEnvironment = KeGetCurrentThread()->ApcStateIndex;
|
|
Irp->Tail.Overlay.CurrentStackLocation = (PIO_STACK_LOCATION)(Irp + 1) + StackSize;
|
|
|
|
/* Initialize the Thread List */
|
|
InitializeListHead(&Irp->ThreadListEntry);
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
BOOLEAN
|
|
NTAPI
|
|
IoIsOperationSynchronous(IN PIRP Irp)
|
|
{
|
|
BOOLEAN SynchIO;
|
|
BOOLEAN ForceAsync;
|
|
|
|
/* If the IRP requests synchronous paging I/O, if the file object was opened
|
|
for synchronous I/O, if the IRP_SYNCHRONOUS_API flag is set in the IRP
|
|
the operation is synchronous */
|
|
SynchIO = (IoGetCurrentIrpStackLocation(Irp)->FileObject->Flags & FO_SYNCHRONOUS_IO) ||
|
|
(Irp->Flags & IRP_SYNCHRONOUS_API) || (Irp->Flags & IRP_SYNCHRONOUS_PAGING_IO);
|
|
|
|
/* If the IRP requests asynchronous paging I/O, the operation is asynchronous,
|
|
even if one of the above conditions is true */
|
|
ForceAsync = (Irp->Flags & IRP_PAGING_IO) && !(Irp->Flags & IRP_SYNCHRONOUS_PAGING_IO);
|
|
|
|
/* Check the flags */
|
|
if (SynchIO && !ForceAsync)
|
|
{
|
|
/* Synch API or Paging I/O is OK, as is Sync File I/O */
|
|
return TRUE;
|
|
}
|
|
|
|
/* Otherwise, it is an asynchronous operation. */
|
|
return FALSE;
|
|
}
|
|
|
|
/*
|
|
* @unimplemented
|
|
*/
|
|
BOOLEAN
|
|
NTAPI
|
|
IoIsValidNameGraftingBuffer(IN PIRP Irp,
|
|
IN PREPARSE_DATA_BUFFER ReparseBuffer)
|
|
{
|
|
UNIMPLEMENTED;
|
|
return FALSE;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
PIRP
|
|
NTAPI
|
|
IoMakeAssociatedIrp(IN PIRP Irp,
|
|
IN CCHAR StackSize)
|
|
{
|
|
PIRP AssocIrp;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Associating IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
|
|
/* Allocate the IRP */
|
|
AssocIrp = IoAllocateIrp(StackSize, FALSE);
|
|
if (!AssocIrp) return NULL;
|
|
|
|
/* Set the Flags */
|
|
AssocIrp->Flags |= IRP_ASSOCIATED_IRP;
|
|
|
|
/* Set the Thread */
|
|
AssocIrp->Tail.Overlay.Thread = Irp->Tail.Overlay.Thread;
|
|
|
|
/* Associate them */
|
|
AssocIrp->AssociatedIrp.MasterIrp = Irp;
|
|
return AssocIrp;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoQueueThreadIrp(IN PIRP Irp)
|
|
{
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Queueing IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
|
|
/* Use our inlined routine */
|
|
IopQueueIrpToThread(Irp);
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
* Reference: Chris Cant's "Writing WDM Device Drivers"
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoReuseIrp(IN OUT PIRP Irp,
|
|
IN NTSTATUS Status)
|
|
{
|
|
UCHAR AllocationFlags;
|
|
IOTRACE(IO_IRP_DEBUG,
|
|
"%s - Reusing IRP %p\n",
|
|
__FUNCTION__,
|
|
Irp);
|
|
|
|
/* Make sure it's OK to reuse it */
|
|
ASSERT(!Irp->CancelRoutine);
|
|
ASSERT(IsListEmpty(&Irp->ThreadListEntry));
|
|
|
|
/* Get the old flags */
|
|
AllocationFlags = Irp->AllocationFlags;
|
|
|
|
/* Reinitialize the IRP */
|
|
IoInitializeIrp(Irp, Irp->Size, Irp->StackCount);
|
|
|
|
/* Duplicate the data */
|
|
Irp->IoStatus.Status = Status;
|
|
Irp->AllocationFlags = AllocationFlags;
|
|
}
|
|
|
|
/*
|
|
* @implemented
|
|
*/
|
|
VOID
|
|
NTAPI
|
|
IoSetTopLevelIrp(IN PIRP Irp)
|
|
{
|
|
/* Set the IRP */
|
|
PsGetCurrentThread()->TopLevelIrp = (ULONG_PTR)Irp;
|
|
}
|
|
|
|
#if defined (_WIN64)
|
|
BOOLEAN
|
|
NTAPI
|
|
IoIs32bitProcess(
|
|
IN PIRP Irp OPTIONAL)
|
|
{
|
|
UNIMPLEMENTED;
|
|
return FALSE;
|
|
}
|
|
#endif
|