mirror of
https://github.com/gnif/LookingGlass.git
synced 2024-11-10 08:38:20 +00:00
956 lines
23 KiB
C++
956 lines
23 KiB
C++
/*
|
|
Looking Glass - KVM FrameRelay (KVMFR) Client
|
|
Copyright (C) 2017 Geoffrey McRae <geoff@hostfission.com>
|
|
https://looking-glass.hostfission.com
|
|
|
|
This program is free software; you can redistribute it and/or modify it under
|
|
the terms of the GNU General Public License as published by the Free Software
|
|
Foundation; either version 2 of the License, or (at your option) any later
|
|
version.
|
|
|
|
This program is distributed in the hope that it will be useful, but WITHOUT ANY
|
|
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A
|
|
PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License along with
|
|
this program; if not, write to the Free Software Foundation, Inc., 59 Temple
|
|
Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include "Capture/DXGI.h"
|
|
using namespace Capture;
|
|
|
|
#include "common/debug.h"
|
|
#include "TraceUtil.h"
|
|
|
|
#include <mfapi.h>
|
|
#include <wmcodecdsp.h>
|
|
#include <codecapi.h>
|
|
#include <mferror.h>
|
|
#include <evr.h>
|
|
#include <mfreadwrite.h>
|
|
|
|
template <class T> void SafeRelease(T **ppT)
|
|
{
|
|
if (*ppT)
|
|
{
|
|
(*ppT)->Release();
|
|
*ppT = NULL;
|
|
}
|
|
}
|
|
|
|
DXGI::DXGI() :
|
|
m_cRef(1),
|
|
m_options(NULL),
|
|
m_initialized(false),
|
|
m_dxgiFactory(),
|
|
m_device(),
|
|
m_deviceContext(),
|
|
m_dup(),
|
|
m_texture(),
|
|
m_pointer(NULL)
|
|
{
|
|
MFStartup(MF_VERSION);
|
|
}
|
|
|
|
DXGI::~DXGI()
|
|
{
|
|
}
|
|
|
|
bool DXGI::Initialize(CaptureOptions * options)
|
|
{
|
|
if (m_initialized)
|
|
DeInitialize();
|
|
|
|
m_options = options;
|
|
HRESULT status;
|
|
|
|
status = CreateDXGIFactory1(__uuidof(IDXGIFactory1), (void **)(&m_dxgiFactory));
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_ERROR("Failed to create DXGIFactory: %08x", (int)status);
|
|
return false;
|
|
}
|
|
|
|
bool done = false;
|
|
IDXGIAdapter1Ptr adapter;
|
|
for (int i = 0; m_dxgiFactory->EnumAdapters1(i, &adapter) != DXGI_ERROR_NOT_FOUND; i++)
|
|
{
|
|
IDXGIOutputPtr output;
|
|
for (int i = 0; adapter->EnumOutputs(i, &output) != DXGI_ERROR_NOT_FOUND; i++)
|
|
{
|
|
DXGI_OUTPUT_DESC outputDesc;
|
|
output->GetDesc(&outputDesc);
|
|
if (!outputDesc.AttachedToDesktop)
|
|
{
|
|
SafeRelease(&output);
|
|
continue;
|
|
}
|
|
|
|
m_output = output;
|
|
if (!m_output)
|
|
{
|
|
SafeRelease(&output);
|
|
SafeRelease(&adapter);
|
|
DEBUG_ERROR("Failed to get IDXGIOutput1");
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
|
|
m_width = outputDesc.DesktopCoordinates.right - outputDesc.DesktopCoordinates.left;
|
|
m_height = outputDesc.DesktopCoordinates.bottom - outputDesc.DesktopCoordinates.top;
|
|
|
|
SafeRelease(&output);
|
|
done = true;
|
|
break;
|
|
}
|
|
|
|
if (done)
|
|
break;
|
|
|
|
SafeRelease(&adapter);
|
|
}
|
|
|
|
if (!done)
|
|
{
|
|
DEBUG_ERROR("Failed to locate a valid output device");
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
|
|
static const D3D_FEATURE_LEVEL featureLevels[] = {
|
|
D3D_FEATURE_LEVEL_11_1,
|
|
D3D_FEATURE_LEVEL_11_0,
|
|
D3D_FEATURE_LEVEL_10_1,
|
|
D3D_FEATURE_LEVEL_10_0,
|
|
D3D_FEATURE_LEVEL_9_3,
|
|
D3D_FEATURE_LEVEL_9_2,
|
|
D3D_FEATURE_LEVEL_9_1
|
|
};
|
|
|
|
#if DEBUG
|
|
#define CREATE_FLAGS (D3D11_CREATE_DEVICE_DEBUG)
|
|
#else
|
|
#define CREATE_FLAGS (0)
|
|
#endif
|
|
|
|
status = D3D11CreateDevice(
|
|
adapter,
|
|
D3D_DRIVER_TYPE_UNKNOWN,
|
|
NULL,
|
|
CREATE_FLAGS | D3D11_CREATE_DEVICE_VIDEO_SUPPORT,
|
|
featureLevels, ARRAYSIZE(featureLevels),
|
|
D3D11_SDK_VERSION,
|
|
&m_device,
|
|
&m_featureLevel,
|
|
&m_deviceContext
|
|
);
|
|
SafeRelease(&adapter);
|
|
#undef CREATE_FLAGS
|
|
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to create D3D11 device", status);
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
|
|
bool h264 = false;
|
|
for(CaptureOptions::const_iterator it = m_options->cbegin(); it != m_options->cend(); ++it)
|
|
{
|
|
if (_stricmp(*it, "h264") == 0) h264 = true;
|
|
}
|
|
|
|
if (h264)
|
|
{
|
|
DEBUG_WARN("Enabling experimental H.264 compression");
|
|
m_frameType = FRAME_TYPE_H264;
|
|
if (!InitH264Capture())
|
|
{
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
m_frameType = FRAME_TYPE_ARGB;
|
|
if (!InitRawCapture())
|
|
{
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
}
|
|
|
|
IDXGIDevicePtr dxgi;
|
|
status = m_device->QueryInterface(__uuidof(IDXGIDevice), (void **)&dxgi);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to obtain the IDXGIDevice interface from the D3D11 device", status);
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
|
|
dxgi->SetGPUThreadPriority(7);
|
|
|
|
// we try this twice just incase we still get an error
|
|
// on re-initialization
|
|
for(int i = 0; i < 2; ++i)
|
|
{
|
|
status = m_output->DuplicateOutput(m_device, &m_dup);
|
|
if (SUCCEEDED(status))
|
|
break;
|
|
Sleep(200);
|
|
}
|
|
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("DuplicateOutput Failed", status);
|
|
DeInitialize();
|
|
return false;
|
|
}
|
|
|
|
m_initialized = true;
|
|
return true;
|
|
}
|
|
|
|
bool DXGI::InitRawCapture()
|
|
{
|
|
D3D11_TEXTURE2D_DESC texDesc;
|
|
ZeroMemory(&texDesc, sizeof(texDesc));
|
|
texDesc.Width = m_width;
|
|
texDesc.Height = m_height;
|
|
texDesc.MipLevels = 1;
|
|
texDesc.ArraySize = 1;
|
|
texDesc.SampleDesc.Count = 1;
|
|
texDesc.SampleDesc.Quality = 0;
|
|
texDesc.Usage = D3D11_USAGE_STAGING;
|
|
texDesc.Format = DXGI_FORMAT_B8G8R8A8_UNORM;
|
|
texDesc.BindFlags = 0;
|
|
texDesc.CPUAccessFlags = D3D11_CPU_ACCESS_READ;
|
|
texDesc.MiscFlags = 0;
|
|
|
|
HRESULT status = m_device->CreateTexture2D(&texDesc, NULL, &m_texture);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to create texture", status);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool DXGI::InitH264Capture()
|
|
{
|
|
HRESULT status;
|
|
|
|
MFT_REGISTER_TYPE_INFO typeInfo;
|
|
IMFActivate **activationPointers;
|
|
UINT32 activationPointerCount;
|
|
|
|
ID3D10MultithreadPtr mt(m_device);
|
|
mt->SetMultithreadProtected(TRUE);
|
|
SafeRelease(&mt);
|
|
|
|
m_encodeEvent = CreateEvent(NULL, TRUE , FALSE, NULL);
|
|
m_shutdownEvent = CreateEvent(NULL, FALSE, FALSE, NULL);
|
|
InitializeCriticalSection(&m_encodeCS);
|
|
|
|
typeInfo.guidMajorType = MFMediaType_Video;
|
|
typeInfo.guidSubtype = MFVideoFormat_H264;
|
|
|
|
status = MFTEnumEx(
|
|
MFT_CATEGORY_VIDEO_ENCODER,
|
|
MFT_ENUM_FLAG_HARDWARE,
|
|
NULL,
|
|
&typeInfo,
|
|
&activationPointers,
|
|
&activationPointerCount
|
|
);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to enumerate encoder MFTs", status);
|
|
return false;
|
|
}
|
|
|
|
if (activationPointerCount == 0)
|
|
{
|
|
DEBUG_WINERROR("Hardware H264 MFT not available", status);
|
|
return false;
|
|
}
|
|
|
|
{
|
|
UINT32 nameLen = 0;
|
|
activationPointers[0]->GetStringLength(MFT_FRIENDLY_NAME_Attribute, &nameLen);
|
|
wchar_t * name = new wchar_t[nameLen+1];
|
|
activationPointers[0]->GetString(MFT_FRIENDLY_NAME_Attribute, name, nameLen + 1, NULL);
|
|
DEBUG_INFO("Using Encoder: %S", name);
|
|
delete[] name;
|
|
}
|
|
|
|
m_mfActivation = activationPointers[0];
|
|
CoTaskMemFree(activationPointers);
|
|
|
|
status = m_mfActivation->ActivateObject(IID_PPV_ARGS(&m_mfTransform));
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to create H264 encoder MFT", status);
|
|
return false;
|
|
}
|
|
|
|
IMFAttributesPtr attribs;
|
|
m_mfTransform->GetAttributes(&attribs);
|
|
attribs->SetUINT32 (MF_READWRITE_ENABLE_HARDWARE_TRANSFORMS , TRUE);
|
|
attribs->SetUINT32 (MF_SOURCE_READER_ENABLE_VIDEO_PROCESSING , TRUE);
|
|
attribs->SetUINT32 (MF_SOURCE_READER_ENABLE_ADVANCED_VIDEO_PROCESSING, TRUE);
|
|
attribs->SetUINT32 (MF_LOW_LATENCY , TRUE);
|
|
|
|
UINT32 d3d11Aware = 0;
|
|
UINT32 async = 0;
|
|
attribs->GetUINT32(MF_TRANSFORM_ASYNC, &async);
|
|
attribs->GetUINT32(MF_SA_D3D11_AWARE, &d3d11Aware);
|
|
if (async)
|
|
attribs->SetUINT32(MF_TRANSFORM_ASYNC_UNLOCK, TRUE);
|
|
SafeRelease(&attribs);
|
|
|
|
status = m_mfTransform.QueryInterface(IID_PPV_ARGS(&m_mediaEventGen));
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to obtain th emedia event generator interface", status);
|
|
return false;
|
|
}
|
|
|
|
status = m_mediaEventGen->BeginGetEvent(this, NULL);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to set the begin get event", status);
|
|
return false;
|
|
}
|
|
|
|
if (d3d11Aware)
|
|
{
|
|
MFCreateDXGIDeviceManager(&m_resetToken, &m_mfDeviceManager);
|
|
status = m_mfDeviceManager->ResetDevice(m_device, m_resetToken);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to call reset device", status);
|
|
return false;
|
|
}
|
|
|
|
status = m_mfTransform->ProcessMessage(MFT_MESSAGE_SET_D3D_MANAGER, ULONG_PTR(m_mfDeviceManager.GetInterfacePtr()));
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to set the D3D manager", status);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
IMFMediaTypePtr outType;
|
|
MFCreateMediaType(&outType);
|
|
|
|
outType->SetGUID (MF_MT_MAJOR_TYPE , MFMediaType_Video);
|
|
outType->SetGUID (MF_MT_SUBTYPE , MFVideoFormat_H264);
|
|
outType->SetUINT32(MF_MT_AVG_BITRATE , 384*1000);
|
|
outType->SetUINT32(MF_MT_INTERLACE_MODE , MFVideoInterlace_Progressive);
|
|
outType->SetUINT32(MF_MT_MPEG2_PROFILE , eAVEncH264VProfile_High);
|
|
outType->SetUINT32(MF_MT_ALL_SAMPLES_INDEPENDENT, TRUE);
|
|
|
|
MFSetAttributeSize (outType, MF_MT_FRAME_SIZE , m_width, m_height);
|
|
MFSetAttributeRatio(outType, MF_MT_FRAME_RATE , 30, 1);
|
|
MFSetAttributeRatio(outType, MF_MT_PIXEL_ASPECT_RATIO, 1, 1);
|
|
|
|
status = m_mfTransform->SetOutputType(0, outType, 0);
|
|
SafeRelease(&outType);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to set the output media type on the H264 encoder MFT", status);
|
|
return false;
|
|
}
|
|
|
|
IMFMediaTypePtr inType;
|
|
MFCreateMediaType(&inType);
|
|
|
|
inType->SetGUID (MF_MT_MAJOR_TYPE , MFMediaType_Video );
|
|
inType->SetGUID (MF_MT_SUBTYPE , MFVideoFormat_NV12);
|
|
inType->SetUINT32(MF_MT_INTERLACE_MODE , MFVideoInterlace_Progressive);
|
|
inType->SetUINT32(MF_MT_ALL_SAMPLES_INDEPENDENT, TRUE);
|
|
|
|
MFSetAttributeSize (inType, MF_MT_FRAME_SIZE , m_width, m_height);
|
|
MFSetAttributeRatio(inType, MF_MT_FRAME_RATE , 30, 1);
|
|
MFSetAttributeRatio(inType, MF_MT_PIXEL_ASPECT_RATIO, 1 , 1);
|
|
|
|
status = m_mfTransform->SetInputType(0, inType, 0);
|
|
SafeRelease(&inType);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to set the input media type on the H264 encoder MFT", status);
|
|
return false;
|
|
}
|
|
|
|
m_mfTransform->ProcessMessage(MFT_MESSAGE_COMMAND_FLUSH , NULL);
|
|
m_mfTransform->ProcessMessage(MFT_MESSAGE_NOTIFY_BEGIN_STREAMING, NULL);
|
|
m_mfTransform->ProcessMessage(MFT_MESSAGE_NOTIFY_START_OF_STREAM, NULL);
|
|
|
|
#if 0
|
|
status = MFTRegisterLocalByCLSID(
|
|
__uuidof(CColorConvertDMO),
|
|
MFT_CATEGORY_VIDEO_PROCESSOR,
|
|
L"",
|
|
MFT_ENUM_FLAG_SYNCMFT,
|
|
0,
|
|
NULL,
|
|
0,
|
|
NULL
|
|
);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_ERROR("Failed to register color converter DSP");
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
return true;
|
|
}
|
|
|
|
void DXGI::DeInitialize()
|
|
{
|
|
if (m_mediaEventGen)
|
|
{
|
|
m_mfTransform->ProcessMessage(MFT_MESSAGE_NOTIFY_END_OF_STREAM, NULL);
|
|
m_mfTransform->ProcessMessage(MFT_MESSAGE_COMMAND_DRAIN, NULL);
|
|
while (WaitForSingleObject(m_shutdownEvent, INFINITE) != WAIT_OBJECT_0) {}
|
|
m_mfTransform->DeleteInputStream(0);
|
|
}
|
|
|
|
if (m_releaseFrame)
|
|
{
|
|
m_releaseFrame = false;
|
|
m_dup->ReleaseFrame();
|
|
}
|
|
|
|
if (m_pointer)
|
|
{
|
|
delete[] m_pointer;
|
|
m_pointer = NULL;
|
|
m_pointerBufSize = 0;
|
|
}
|
|
|
|
if (m_surfaceMapped)
|
|
{
|
|
m_deviceContext->Unmap(m_texture, 0);
|
|
m_surfaceMapped = false;
|
|
}
|
|
|
|
SafeRelease(&m_mediaEventGen);
|
|
SafeRelease(&m_mfTransform);
|
|
SafeRelease(&m_mfDeviceManager);
|
|
|
|
SafeRelease(&m_texture);
|
|
SafeRelease(&m_dup);
|
|
SafeRelease(&m_output);
|
|
SafeRelease(&m_deviceContext);
|
|
SafeRelease(&m_device);
|
|
SafeRelease(&m_dxgiFactory);
|
|
|
|
if (m_encodeEvent)
|
|
{
|
|
CloseHandle(m_encodeEvent );
|
|
CloseHandle(m_shutdownEvent);
|
|
m_encodeEvent = NULL;
|
|
m_shutdownEvent = NULL;
|
|
DeleteCriticalSection(&m_encodeCS);
|
|
}
|
|
|
|
if (m_mfActivation)
|
|
{
|
|
m_mfActivation->ShutdownObject();
|
|
SafeRelease(&m_mfActivation);
|
|
}
|
|
|
|
m_initialized = false;
|
|
}
|
|
|
|
FrameType DXGI::GetFrameType()
|
|
{
|
|
if (!m_initialized)
|
|
return FRAME_TYPE_INVALID;
|
|
return m_frameType;
|
|
}
|
|
|
|
size_t DXGI::GetMaxFrameSize()
|
|
{
|
|
if (!m_initialized)
|
|
return 0;
|
|
|
|
return (m_width * m_height * 4);
|
|
}
|
|
|
|
STDMETHODIMP Capture::DXGI::Invoke(IMFAsyncResult * pAsyncResult)
|
|
{
|
|
HRESULT status, evtStatus;
|
|
MediaEventType meType = MEUnknown;
|
|
IMFMediaEvent *pEvent = NULL;
|
|
|
|
status = m_mediaEventGen->EndGetEvent(pAsyncResult, &pEvent);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("EndGetEvent", status);
|
|
return status;
|
|
}
|
|
|
|
status = pEvent->GetStatus(&evtStatus);
|
|
if (FAILED(status))
|
|
{
|
|
SafeRelease(&pEvent);
|
|
DEBUG_WINERROR("GetStatus", status);
|
|
return status;
|
|
}
|
|
|
|
if (FAILED(evtStatus))
|
|
{
|
|
SafeRelease(&pEvent);
|
|
DEBUG_WINERROR("evtStatus", evtStatus);
|
|
return evtStatus;
|
|
}
|
|
|
|
status = pEvent->GetType(&meType);
|
|
if (FAILED(status))
|
|
{
|
|
SafeRelease(&pEvent);
|
|
DEBUG_WINERROR("GetType", status);
|
|
return status;
|
|
}
|
|
SafeRelease(&pEvent);
|
|
|
|
switch (meType)
|
|
{
|
|
case METransformNeedInput:
|
|
EnterCriticalSection(&m_encodeCS);
|
|
m_encodeNeedsData = true;
|
|
SetEvent(m_encodeEvent);
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
break;
|
|
|
|
case METransformHaveOutput:
|
|
EnterCriticalSection(&m_encodeCS);
|
|
m_encodeHasData = true;
|
|
SetEvent(m_encodeEvent);
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
break;
|
|
|
|
case METransformDrainComplete:
|
|
{
|
|
status = m_mfTransform->ProcessMessage(MFT_MESSAGE_COMMAND_FLUSH, 0);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("MFT_MESSAGE_COMMAND_FLUSH", status);
|
|
return status;
|
|
}
|
|
|
|
SetEvent(m_shutdownEvent);
|
|
return S_OK;
|
|
}
|
|
|
|
case MEError:
|
|
DEBUG_INFO("err");
|
|
break;
|
|
|
|
default:
|
|
DEBUG_INFO("unk");
|
|
break;
|
|
}
|
|
|
|
status = m_mediaEventGen->BeginGetEvent(this, NULL);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("BeginGetEvent", status);
|
|
return status;
|
|
}
|
|
|
|
return status;
|
|
}
|
|
|
|
void DXGI::WaitForDesktop()
|
|
{
|
|
HDESK desktop;
|
|
do
|
|
{
|
|
desktop = OpenInputDesktop(0, TRUE, GENERIC_READ);
|
|
if (desktop)
|
|
break;
|
|
Sleep(100);
|
|
}
|
|
while (!desktop);
|
|
CloseDesktop(desktop);
|
|
}
|
|
|
|
GrabStatus Capture::DXGI::GrabFrameTexture(FrameInfo & frame, ID3D11Texture2DPtr & texture, bool & timeout)
|
|
{
|
|
if (!m_initialized)
|
|
return GRAB_STATUS_ERROR;
|
|
|
|
timeout = false;
|
|
DXGI_OUTDUPL_FRAME_INFO frameInfo;
|
|
IDXGIResourcePtr res;
|
|
|
|
HRESULT status;
|
|
bool cursorUpdate = false;
|
|
for (int i = 0; i < 2; ++i)
|
|
{
|
|
while (true)
|
|
{
|
|
if (m_releaseFrame)
|
|
{
|
|
m_releaseFrame = false;
|
|
status = m_dup->ReleaseFrame();
|
|
|
|
switch (status)
|
|
{
|
|
case S_OK:
|
|
break;
|
|
|
|
case DXGI_ERROR_INVALID_CALL:
|
|
DEBUG_ERROR("Frame was already released");
|
|
return GRAB_STATUS_ERROR;
|
|
|
|
case DXGI_ERROR_ACCESS_LOST:
|
|
WaitForDesktop();
|
|
return GRAB_STATUS_REINIT;
|
|
}
|
|
}
|
|
|
|
status = m_dup->AcquireNextFrame(1000, &frameInfo, &res);
|
|
if (status == DXGI_ERROR_WAIT_TIMEOUT)
|
|
{
|
|
timeout = true;
|
|
return GRAB_STATUS_OK;
|
|
}
|
|
|
|
if (!SUCCEEDED(status))
|
|
break;
|
|
|
|
m_releaseFrame = true;
|
|
|
|
// if we have a mouse update
|
|
if (frameInfo.LastMouseUpdateTime.QuadPart)
|
|
{
|
|
if (
|
|
m_lastMousePos.x != frameInfo.PointerPosition.Position.x ||
|
|
m_lastMousePos.y != frameInfo.PointerPosition.Position.y
|
|
) {
|
|
cursorUpdate = true;
|
|
frame.cursor.hasPos = true;
|
|
frame.cursor.x = frameInfo.PointerPosition.Position.x;
|
|
frame.cursor.y = frameInfo.PointerPosition.Position.y;
|
|
m_lastMousePos.x = frameInfo.PointerPosition.Position.x;
|
|
m_lastMousePos.y = frameInfo.PointerPosition.Position.y;
|
|
}
|
|
|
|
if (m_lastMouseVis != frameInfo.PointerPosition.Visible)
|
|
{
|
|
cursorUpdate = true;
|
|
m_lastMouseVis = frameInfo.PointerPosition.Visible;
|
|
}
|
|
|
|
frame.cursor.visible = m_lastMouseVis == TRUE;
|
|
}
|
|
|
|
// if the pointer shape has changed
|
|
if (frameInfo.PointerShapeBufferSize > 0)
|
|
{
|
|
cursorUpdate = true;
|
|
if (m_pointerBufSize < frameInfo.PointerShapeBufferSize)
|
|
{
|
|
if (m_pointer)
|
|
delete[] m_pointer;
|
|
m_pointer = new BYTE[frameInfo.PointerShapeBufferSize];
|
|
m_pointerBufSize = frameInfo.PointerShapeBufferSize;
|
|
}
|
|
|
|
DXGI_OUTDUPL_POINTER_SHAPE_INFO shapeInfo;
|
|
status = m_dup->GetFramePointerShape(m_pointerBufSize, m_pointer, &m_pointerSize, &shapeInfo);
|
|
if (!SUCCEEDED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to get the new pointer shape", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
switch (shapeInfo.Type)
|
|
{
|
|
case DXGI_OUTDUPL_POINTER_SHAPE_TYPE_COLOR : frame.cursor.type = CURSOR_TYPE_COLOR; break;
|
|
case DXGI_OUTDUPL_POINTER_SHAPE_TYPE_MASKED_COLOR: frame.cursor.type = CURSOR_TYPE_MASKED_COLOR; break;
|
|
case DXGI_OUTDUPL_POINTER_SHAPE_TYPE_MONOCHROME : frame.cursor.type = CURSOR_TYPE_MONOCHROME; break;
|
|
default:
|
|
DEBUG_ERROR("Invalid cursor type");
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
frame.cursor.hasShape = true;
|
|
frame.cursor.shape = m_pointer;
|
|
frame.cursor.w = shapeInfo.Width;
|
|
frame.cursor.h = shapeInfo.Height;
|
|
frame.cursor.pitch = shapeInfo.Pitch;
|
|
frame.cursor.dataSize = m_pointerSize;
|
|
}
|
|
|
|
if (frameInfo.LastPresentTime.QuadPart != 0)
|
|
break;
|
|
|
|
SafeRelease(&res);
|
|
|
|
if (cursorUpdate)
|
|
return GRAB_STATUS_CURSOR;
|
|
}
|
|
|
|
if (SUCCEEDED(status))
|
|
break;
|
|
|
|
switch (status)
|
|
{
|
|
// desktop switch, mode change, switch DWM on or off or Secure Desktop
|
|
case DXGI_ERROR_ACCESS_LOST:
|
|
case WAIT_ABANDONED:
|
|
WaitForDesktop();
|
|
return GRAB_STATUS_REINIT;
|
|
|
|
default:
|
|
// unknown failure
|
|
DEBUG_WINERROR("AcquireNextFrame failed", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
}
|
|
|
|
// retry count exceeded
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to acquire next frame", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
res.QueryInterface(IID_PPV_ARGS(&texture));
|
|
SafeRelease(&res);
|
|
|
|
if (!texture)
|
|
{
|
|
DEBUG_ERROR("Failed to get src ID3D11Texture2D");
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
return GRAB_STATUS_OK;
|
|
}
|
|
|
|
GrabStatus Capture::DXGI::GrabFrameRaw(FrameInfo & frame)
|
|
{
|
|
GrabStatus result;
|
|
ID3D11Texture2DPtr src;
|
|
bool timeout;
|
|
|
|
while(true)
|
|
{
|
|
result = GrabFrameTexture(frame, src, timeout);
|
|
if (result != GRAB_STATUS_OK)
|
|
return result;
|
|
|
|
if (timeout)
|
|
{
|
|
if (!m_surfaceMapped)
|
|
continue;
|
|
m_memcpy.Wake();
|
|
|
|
// send the last frame again if we timeout to prevent the client stalling on restart
|
|
frame.pitch = m_mapping.RowPitch;
|
|
frame.stride = m_mapping.RowPitch / 4;
|
|
|
|
unsigned int size = m_height * m_mapping.RowPitch;
|
|
m_memcpy.Copy(frame.buffer, m_mapping.pData, size < frame.bufferSize ? size : frame.bufferSize);
|
|
return GRAB_STATUS_OK;
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
m_deviceContext->CopyResource(m_texture, src);
|
|
SafeRelease(&src);
|
|
|
|
if (m_surfaceMapped)
|
|
{
|
|
m_deviceContext->Unmap(m_texture, 0);
|
|
m_surfaceMapped = false;
|
|
}
|
|
|
|
HRESULT status;
|
|
status = m_deviceContext->Map(m_texture, 0, D3D11_MAP_READ, 0, &m_mapping);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to map the texture", status);
|
|
DeInitialize();
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
m_surfaceMapped = true;
|
|
|
|
TRACE_START("DXGI Memory Copy");
|
|
// wake up the copy threads
|
|
m_memcpy.Wake();
|
|
|
|
frame.pitch = m_mapping.RowPitch;
|
|
frame.stride = m_mapping.RowPitch >> 2;
|
|
|
|
const unsigned int size = m_height * m_mapping.RowPitch;
|
|
m_memcpy.Copy(frame.buffer, m_mapping.pData, size < frame.bufferSize ? size : frame.bufferSize);
|
|
TRACE_END;
|
|
|
|
return GRAB_STATUS_OK;
|
|
}
|
|
|
|
GrabStatus Capture::DXGI::GrabFrameH264(FrameInfo & frame)
|
|
{
|
|
while(true)
|
|
{
|
|
// only reset the event if there isn't work pending
|
|
EnterCriticalSection(&m_encodeCS);
|
|
if (!m_encodeHasData && !m_encodeNeedsData)
|
|
ResetEvent(m_encodeEvent);
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
|
|
switch (WaitForSingleObject(m_encodeEvent, 1000))
|
|
{
|
|
case WAIT_FAILED:
|
|
DEBUG_WINERROR("Wait for encode event failed", GetLastError());
|
|
return GRAB_STATUS_ERROR;
|
|
|
|
case WAIT_ABANDONED:
|
|
DEBUG_ERROR("Wait abandoned");
|
|
return GRAB_STATUS_ERROR;
|
|
|
|
case WAIT_TIMEOUT:
|
|
continue;
|
|
|
|
case WAIT_OBJECT_0:
|
|
break;
|
|
}
|
|
|
|
EnterCriticalSection(&m_encodeCS);
|
|
|
|
HRESULT status;
|
|
if (m_encodeNeedsData)
|
|
{
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
GrabStatus result;
|
|
ID3D11Texture2DPtr src;
|
|
bool timeout;
|
|
|
|
while(true)
|
|
{
|
|
result = GrabFrameTexture(frame, src, timeout);
|
|
if (result != GRAB_STATUS_OK)
|
|
{
|
|
return result;
|
|
}
|
|
|
|
//FIXME: we should send the last frame again
|
|
if (!timeout)
|
|
break;
|
|
}
|
|
|
|
// cursor data may be returned, only turn off the flag if we have a frame
|
|
EnterCriticalSection(&m_encodeCS);
|
|
m_encodeNeedsData = false;
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
|
|
IMFMediaBufferPtr buffer;
|
|
status = MFCreateDXGISurfaceBuffer(__uuidof(ID3D11Texture2D), src, 0, FALSE, &buffer);
|
|
SafeRelease(&src);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to create DXGI surface buffer from texture", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
IMF2DBufferPtr imfBuffer(buffer);
|
|
DWORD length;
|
|
imfBuffer->GetContiguousLength(&length);
|
|
buffer->SetCurrentLength(length);
|
|
SafeRelease(&imfBuffer);
|
|
|
|
IMFSamplePtr sample;
|
|
MFCreateSample(&sample);
|
|
sample->AddBuffer(buffer);
|
|
|
|
status = m_mfTransform->ProcessInput(0, sample, 0);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("Failed to process the input", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
SafeRelease(&src );
|
|
SafeRelease(&sample);
|
|
SafeRelease(&buffer);
|
|
|
|
EnterCriticalSection(&m_encodeCS);
|
|
}
|
|
|
|
if (m_encodeHasData)
|
|
{
|
|
m_encodeHasData = false;
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
|
|
// wake up the copy threads
|
|
TRACE_START("copy");
|
|
m_memcpy.Wake();
|
|
|
|
MFT_OUTPUT_STREAM_INFO streamInfo;
|
|
status = m_mfTransform->GetOutputStreamInfo(0, &streamInfo);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("GetOutputStreamInfo", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
DWORD outStatus;
|
|
MFT_OUTPUT_DATA_BUFFER outDataBuffer;
|
|
outDataBuffer.dwStreamID = 0;
|
|
outDataBuffer.dwStatus = 0;
|
|
outDataBuffer.pEvents = NULL;
|
|
outDataBuffer.pSample = NULL;
|
|
|
|
status = m_mfTransform->ProcessOutput(0, 1, &outDataBuffer, &outStatus);
|
|
if (FAILED(status))
|
|
{
|
|
DEBUG_WINERROR("ProcessOutput", status);
|
|
return GRAB_STATUS_ERROR;
|
|
}
|
|
|
|
IMFMediaBufferPtr buffer;
|
|
MFCreateAlignedMemoryBuffer((DWORD)frame.bufferSize, MF_128_BYTE_ALIGNMENT, &buffer);
|
|
outDataBuffer.pSample->CopyToBuffer(buffer);
|
|
SafeRelease(&outDataBuffer.pEvents);
|
|
SafeRelease(&outDataBuffer.pSample);
|
|
|
|
BYTE *pixels;
|
|
DWORD maxLen, curLen;
|
|
buffer->Lock(&pixels, &maxLen, &curLen);
|
|
m_memcpy.Copy(frame.buffer, pixels, curLen);
|
|
buffer->Unlock();
|
|
SafeRelease(&buffer);
|
|
|
|
frame.stride = 0;
|
|
frame.pitch = curLen;
|
|
|
|
TRACE_END;
|
|
return GRAB_STATUS_OK;
|
|
}
|
|
|
|
LeaveCriticalSection(&m_encodeCS);
|
|
}
|
|
}
|
|
|
|
GrabStatus DXGI::GrabFrame(FrameInfo & frame)
|
|
{
|
|
frame.width = m_width;
|
|
frame.height = m_height;
|
|
|
|
if (m_frameType == FRAME_TYPE_H264)
|
|
return GrabFrameH264(frame);
|
|
else
|
|
return GrabFrameRaw(frame);
|
|
} |