mirror of
https://github.com/WinampDesktop/winamp.git
synced 2025-04-28 00:25:41 -04:00
1195 lines
49 KiB
C++
1195 lines
49 KiB
C++
#include "gpu_hw_vulkan.h"
|
|
#include "common/assert.h"
|
|
#include "common/log.h"
|
|
#include "common/scope_guard.h"
|
|
#include "common/vulkan/builders.h"
|
|
#include "common/vulkan/context.h"
|
|
#include "common/vulkan/shader_cache.h"
|
|
#include "common/vulkan/util.h"
|
|
#include "gpu_hw_shadergen.h"
|
|
#include "host_display.h"
|
|
#include "host_interface.h"
|
|
#include "system.h"
|
|
Log_SetChannel(GPU_HW_Vulkan);
|
|
|
|
GPU_HW_Vulkan::GPU_HW_Vulkan() = default;
|
|
|
|
GPU_HW_Vulkan::~GPU_HW_Vulkan()
|
|
{
|
|
if (m_host_display)
|
|
{
|
|
m_host_display->ClearDisplayTexture();
|
|
ResetGraphicsAPIState();
|
|
}
|
|
|
|
DestroyResources();
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::Initialize(HostDisplay* host_display, System* system, DMA* dma,
|
|
InterruptController* interrupt_controller, Timers* timers)
|
|
{
|
|
if (host_display->GetRenderAPI() != HostDisplay::RenderAPI::Vulkan)
|
|
{
|
|
Log_ErrorPrintf("Host render API is incompatible");
|
|
return false;
|
|
}
|
|
|
|
Assert(g_vulkan_shader_cache);
|
|
SetCapabilities();
|
|
|
|
if (!GPU_HW::Initialize(host_display, system, dma, interrupt_controller, timers))
|
|
return false;
|
|
|
|
if (!CreatePipelineLayouts())
|
|
{
|
|
Log_ErrorPrintf("Failed to create pipeline layouts");
|
|
return false;
|
|
}
|
|
|
|
if (!CreateSamplers())
|
|
{
|
|
Log_ErrorPrintf("Failed to create samplers");
|
|
return false;
|
|
}
|
|
|
|
if (!CreateVertexBuffer())
|
|
{
|
|
Log_ErrorPrintf("Failed to create vertex buffer");
|
|
return false;
|
|
}
|
|
|
|
if (!CreateUniformBuffer())
|
|
{
|
|
Log_ErrorPrintf("Failed to create uniform buffer");
|
|
return false;
|
|
}
|
|
|
|
if (!CreateTextureBuffer())
|
|
{
|
|
Log_ErrorPrintf("Failed to create texture buffer");
|
|
return false;
|
|
}
|
|
|
|
if (!CreateFramebuffer())
|
|
{
|
|
Log_ErrorPrintf("Failed to create framebuffer");
|
|
return false;
|
|
}
|
|
|
|
if (!CompilePipelines())
|
|
{
|
|
Log_ErrorPrintf("Failed to compile pipelines");
|
|
return false;
|
|
}
|
|
|
|
RestoreGraphicsAPIState();
|
|
return true;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::Reset()
|
|
{
|
|
GPU_HW::Reset();
|
|
|
|
if (InRenderPass())
|
|
EndRenderPass();
|
|
|
|
ClearFramebuffer();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::ResetGraphicsAPIState()
|
|
{
|
|
GPU_HW::ResetGraphicsAPIState();
|
|
|
|
if (InRenderPass())
|
|
EndRenderPass();
|
|
|
|
// vram texture is probably going to be displayed now
|
|
if (!IsDisplayDisabled())
|
|
{
|
|
m_vram_texture.TransitionToLayout(g_vulkan_context->GetCurrentCommandBuffer(),
|
|
VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
}
|
|
}
|
|
|
|
void GPU_HW_Vulkan::RestoreGraphicsAPIState()
|
|
{
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
VkDeviceSize vertex_buffer_offset = 0;
|
|
vkCmdBindVertexBuffers(cmdbuf, 0, 1, m_vertex_stream_buffer.GetBufferPointer(), &vertex_buffer_offset);
|
|
Vulkan::Util::SetViewport(cmdbuf, 0, 0, m_vram_texture.GetWidth(), m_vram_texture.GetHeight());
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_batch_pipeline_layout, 0, 1,
|
|
&m_batch_descriptor_set, 1, &m_current_uniform_buffer_offset);
|
|
SetScissorFromDrawingArea();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UpdateSettings()
|
|
{
|
|
GPU_HW::UpdateSettings();
|
|
|
|
// Everything should be finished executing before recreating resources.
|
|
g_vulkan_context->ExecuteCommandBuffer(true);
|
|
|
|
CreateFramebuffer();
|
|
DestroyPipelines();
|
|
CompilePipelines();
|
|
RestoreGraphicsAPIState();
|
|
UpdateDisplay();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::MapBatchVertexPointer(u32 required_vertices)
|
|
{
|
|
DebugAssert(!m_batch_start_vertex_ptr);
|
|
|
|
const u32 required_space = required_vertices * sizeof(BatchVertex);
|
|
if (!m_vertex_stream_buffer.ReserveMemory(required_space, sizeof(BatchVertex)))
|
|
{
|
|
Log_PerfPrintf("Executing command buffer while waiting for %u bytes in vertex stream buffer", required_space);
|
|
EndRenderPass();
|
|
g_vulkan_context->ExecuteCommandBuffer(false);
|
|
RestoreGraphicsAPIState();
|
|
if (!m_vertex_stream_buffer.ReserveMemory(required_space, sizeof(BatchVertex)))
|
|
Panic("Failed to reserve vertex stream buffer memory");
|
|
}
|
|
|
|
m_batch_start_vertex_ptr = static_cast<BatchVertex*>(m_vertex_stream_buffer.GetCurrentHostPointer());
|
|
m_batch_current_vertex_ptr = m_batch_start_vertex_ptr;
|
|
m_batch_end_vertex_ptr = m_batch_start_vertex_ptr + (m_vertex_stream_buffer.GetCurrentSpace() / sizeof(BatchVertex));
|
|
m_batch_base_vertex = m_vertex_stream_buffer.GetCurrentOffset() / sizeof(BatchVertex);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UnmapBatchVertexPointer(u32 used_vertices)
|
|
{
|
|
DebugAssert(m_batch_start_vertex_ptr);
|
|
if (used_vertices > 0)
|
|
m_vertex_stream_buffer.CommitMemory(used_vertices * sizeof(BatchVertex));
|
|
|
|
m_batch_start_vertex_ptr = nullptr;
|
|
m_batch_end_vertex_ptr = nullptr;
|
|
m_batch_current_vertex_ptr = nullptr;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UploadUniformBuffer(const void* data, u32 data_size)
|
|
{
|
|
const u32 alignment = static_cast<u32>(g_vulkan_context->GetUniformBufferAlignment());
|
|
if (!m_uniform_stream_buffer.ReserveMemory(data_size, alignment))
|
|
{
|
|
Log_PerfPrintf("Executing command buffer while waiting for %u bytes in uniform stream buffer", data_size);
|
|
EndRenderPass();
|
|
g_vulkan_context->ExecuteCommandBuffer(false);
|
|
RestoreGraphicsAPIState();
|
|
if (!m_uniform_stream_buffer.ReserveMemory(data_size, alignment))
|
|
Panic("Failed to reserve uniform stream buffer memory");
|
|
}
|
|
|
|
m_current_uniform_buffer_offset = m_uniform_stream_buffer.GetCurrentOffset();
|
|
std::memcpy(m_uniform_stream_buffer.GetCurrentHostPointer(), data, data_size);
|
|
m_uniform_stream_buffer.CommitMemory(data_size);
|
|
|
|
vkCmdBindDescriptorSets(g_vulkan_context->GetCurrentCommandBuffer(), VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
m_batch_pipeline_layout, 0, 1, &m_batch_descriptor_set, 1, &m_current_uniform_buffer_offset);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::SetCapabilities()
|
|
{
|
|
const u32 max_texture_size = g_vulkan_context->GetDeviceLimits().maxImageDimension2D;
|
|
const u32 max_texture_scale = max_texture_size / VRAM_WIDTH;
|
|
|
|
Log_InfoPrintf("Max texture size: %ux%u", max_texture_size, max_texture_size);
|
|
|
|
m_max_resolution_scale = max_texture_scale;
|
|
m_supports_dual_source_blend = true;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::DestroyResources()
|
|
{
|
|
// Everything should be finished executing before recreating resources.
|
|
if (g_vulkan_context)
|
|
g_vulkan_context->ExecuteCommandBuffer(true);
|
|
|
|
DestroyFramebuffer();
|
|
DestroyPipelines();
|
|
|
|
Vulkan::Util::SafeFreeGlobalDescriptorSet(m_vram_write_descriptor_set);
|
|
Vulkan::Util::SafeDestroyBufferView(m_texture_stream_buffer_view);
|
|
|
|
m_vertex_stream_buffer.Destroy(false);
|
|
m_uniform_stream_buffer.Destroy(false);
|
|
m_texture_stream_buffer.Destroy(false);
|
|
|
|
Vulkan::Util::SafeDestroyPipelineLayout(m_vram_write_pipeline_layout);
|
|
Vulkan::Util::SafeDestroyPipelineLayout(m_single_sampler_pipeline_layout);
|
|
Vulkan::Util::SafeDestroyPipelineLayout(m_no_samplers_pipeline_layout);
|
|
Vulkan::Util::SafeDestroyPipelineLayout(m_batch_pipeline_layout);
|
|
Vulkan::Util::SafeDestroyDescriptorSetLayout(m_vram_write_descriptor_set_layout);
|
|
Vulkan::Util::SafeDestroyDescriptorSetLayout(m_single_sampler_descriptor_set_layout);
|
|
Vulkan::Util::SafeDestroyDescriptorSetLayout(m_batch_descriptor_set_layout);
|
|
Vulkan::Util::SafeDestroySampler(m_point_sampler);
|
|
Vulkan::Util::SafeDestroySampler(m_linear_sampler);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::BeginRenderPass(VkRenderPass render_pass, VkFramebuffer framebuffer, u32 x, u32 y, u32 width,
|
|
u32 height)
|
|
{
|
|
DebugAssert(m_current_render_pass == VK_NULL_HANDLE);
|
|
|
|
const VkRenderPassBeginInfo bi = {VK_STRUCTURE_TYPE_RENDER_PASS_BEGIN_INFO,
|
|
nullptr,
|
|
render_pass,
|
|
framebuffer,
|
|
{{static_cast<s32>(x), static_cast<s32>(y)}, {width, height}},
|
|
0u,
|
|
nullptr};
|
|
vkCmdBeginRenderPass(g_vulkan_context->GetCurrentCommandBuffer(), &bi, VK_SUBPASS_CONTENTS_INLINE);
|
|
m_current_render_pass = render_pass;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::BeginVRAMRenderPass()
|
|
{
|
|
if (m_current_render_pass == m_vram_render_pass)
|
|
return;
|
|
|
|
EndRenderPass();
|
|
BeginRenderPass(m_vram_render_pass, m_vram_framebuffer, 0, 0, m_vram_texture.GetWidth(), m_vram_texture.GetHeight());
|
|
}
|
|
|
|
void GPU_HW_Vulkan::EndRenderPass()
|
|
{
|
|
if (m_current_render_pass == VK_NULL_HANDLE)
|
|
return;
|
|
|
|
vkCmdEndRenderPass(g_vulkan_context->GetCurrentCommandBuffer());
|
|
m_current_render_pass = VK_NULL_HANDLE;
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreatePipelineLayouts()
|
|
{
|
|
VkDevice device = g_vulkan_context->GetDevice();
|
|
|
|
Vulkan::DescriptorSetLayoutBuilder dslbuilder;
|
|
dslbuilder.AddBinding(0, VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC, 1,
|
|
VK_SHADER_STAGE_VERTEX_BIT | VK_SHADER_STAGE_FRAGMENT_BIT);
|
|
dslbuilder.AddBinding(1, VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER, 1, VK_SHADER_STAGE_FRAGMENT_BIT);
|
|
m_batch_descriptor_set_layout = dslbuilder.Create(device);
|
|
if (m_batch_descriptor_set_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
// textures start at 1
|
|
dslbuilder.AddBinding(1, VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER, 1, VK_SHADER_STAGE_FRAGMENT_BIT);
|
|
m_single_sampler_descriptor_set_layout = dslbuilder.Create(device);
|
|
if (m_single_sampler_descriptor_set_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
dslbuilder.AddBinding(0, VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER, 1, VK_SHADER_STAGE_FRAGMENT_BIT);
|
|
m_vram_write_descriptor_set_layout = dslbuilder.Create(device);
|
|
if (m_vram_write_descriptor_set_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
Vulkan::PipelineLayoutBuilder plbuilder;
|
|
plbuilder.AddDescriptorSet(m_batch_descriptor_set_layout);
|
|
m_batch_pipeline_layout = plbuilder.Create(device);
|
|
if (m_batch_pipeline_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
plbuilder.AddDescriptorSet(m_single_sampler_descriptor_set_layout);
|
|
plbuilder.AddPushConstants(VK_SHADER_STAGE_FRAGMENT_BIT, 0, MAX_PUSH_CONSTANTS_SIZE);
|
|
m_single_sampler_pipeline_layout = plbuilder.Create(device);
|
|
if (m_single_sampler_pipeline_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
plbuilder.AddPushConstants(VK_SHADER_STAGE_FRAGMENT_BIT, 0, MAX_PUSH_CONSTANTS_SIZE);
|
|
m_no_samplers_pipeline_layout = plbuilder.Create(device);
|
|
if (m_no_samplers_pipeline_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
plbuilder.AddDescriptorSet(m_vram_write_descriptor_set_layout);
|
|
plbuilder.AddPushConstants(VK_SHADER_STAGE_FRAGMENT_BIT, 0, MAX_PUSH_CONSTANTS_SIZE);
|
|
m_vram_write_pipeline_layout = plbuilder.Create(device);
|
|
if (m_vram_write_pipeline_layout == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreateSamplers()
|
|
{
|
|
VkDevice device = g_vulkan_context->GetDevice();
|
|
|
|
Vulkan::SamplerBuilder sbuilder;
|
|
sbuilder.SetPointSampler(VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER);
|
|
m_point_sampler = sbuilder.Create(device);
|
|
if (m_point_sampler == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
sbuilder.SetLinearSampler(false, VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER);
|
|
m_linear_sampler = sbuilder.Create(device);
|
|
if (m_linear_sampler == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreateFramebuffer()
|
|
{
|
|
// save old vram texture/fbo, in case we're changing scale
|
|
auto old_vram_texture = std::move(m_vram_texture);
|
|
DestroyFramebuffer();
|
|
|
|
// scale vram size to internal resolution
|
|
const u32 texture_width = VRAM_WIDTH * m_resolution_scale;
|
|
const u32 texture_height = VRAM_HEIGHT * m_resolution_scale;
|
|
const VkFormat texture_format = VK_FORMAT_R8G8B8A8_UNORM;
|
|
const VkFormat depth_format = VK_FORMAT_D16_UNORM;
|
|
const VkSampleCountFlagBits samples = VK_SAMPLE_COUNT_1_BIT;
|
|
|
|
if (!m_vram_texture.Create(texture_width, texture_height, 1, 1, texture_format, samples, VK_IMAGE_VIEW_TYPE_2D,
|
|
VK_IMAGE_TILING_OPTIMAL,
|
|
VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_SAMPLED_BIT |
|
|
VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT) ||
|
|
!m_vram_depth_texture.Create(texture_width, texture_height, 1, 1, depth_format, samples, VK_IMAGE_VIEW_TYPE_2D,
|
|
VK_IMAGE_TILING_OPTIMAL,
|
|
VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT) ||
|
|
!m_vram_read_texture.Create(texture_width, texture_height, 1, 1, texture_format, samples, VK_IMAGE_VIEW_TYPE_2D,
|
|
VK_IMAGE_TILING_OPTIMAL,
|
|
VK_IMAGE_USAGE_SAMPLED_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT) ||
|
|
!m_display_texture.Create(texture_width, texture_height, 1, 1, texture_format, samples, VK_IMAGE_VIEW_TYPE_2D,
|
|
VK_IMAGE_TILING_OPTIMAL,
|
|
VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_SAMPLED_BIT |
|
|
VK_IMAGE_USAGE_TRANSFER_SRC_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT) ||
|
|
!m_vram_readback_texture.Create(VRAM_WIDTH, VRAM_HEIGHT, 1, 1, texture_format, samples, VK_IMAGE_VIEW_TYPE_2D,
|
|
VK_IMAGE_TILING_OPTIMAL,
|
|
VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_TRANSFER_SRC_BIT) ||
|
|
!m_vram_readback_staging_texture.Create(Vulkan::StagingBuffer::Type::Readback, texture_format, VRAM_WIDTH,
|
|
VRAM_HEIGHT))
|
|
{
|
|
return false;
|
|
}
|
|
|
|
m_vram_render_pass =
|
|
g_vulkan_context->GetRenderPass(texture_format, depth_format, samples, VK_ATTACHMENT_LOAD_OP_LOAD);
|
|
m_vram_update_depth_render_pass =
|
|
g_vulkan_context->GetRenderPass(VK_FORMAT_UNDEFINED, depth_format, samples, VK_ATTACHMENT_LOAD_OP_DONT_CARE);
|
|
m_display_render_pass = g_vulkan_context->GetRenderPass(m_display_texture.GetFormat(), VK_FORMAT_UNDEFINED,
|
|
m_display_texture.GetSamples(), VK_ATTACHMENT_LOAD_OP_LOAD);
|
|
m_vram_readback_render_pass =
|
|
g_vulkan_context->GetRenderPass(m_vram_readback_texture.GetFormat(), VK_FORMAT_UNDEFINED,
|
|
m_vram_readback_texture.GetSamples(), VK_ATTACHMENT_LOAD_OP_DONT_CARE);
|
|
|
|
if (m_vram_render_pass == VK_NULL_HANDLE || m_vram_update_depth_render_pass == VK_NULL_HANDLE ||
|
|
m_display_render_pass == VK_NULL_HANDLE || m_vram_readback_render_pass == VK_NULL_HANDLE)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
// vram framebuffer has both colour and depth
|
|
{
|
|
Vulkan::FramebufferBuilder fbb;
|
|
fbb.AddAttachment(m_vram_texture.GetView());
|
|
fbb.AddAttachment(m_vram_depth_texture.GetView());
|
|
fbb.SetRenderPass(m_vram_render_pass);
|
|
fbb.SetSize(m_vram_texture.GetWidth(), m_vram_texture.GetHeight(), m_vram_texture.GetLayers());
|
|
m_vram_framebuffer = fbb.Create(g_vulkan_context->GetDevice());
|
|
if (m_vram_framebuffer == VK_NULL_HANDLE)
|
|
return false;
|
|
}
|
|
|
|
m_vram_update_depth_framebuffer = m_vram_depth_texture.CreateFramebuffer(m_vram_update_depth_render_pass);
|
|
m_vram_readback_framebuffer = m_vram_readback_texture.CreateFramebuffer(m_vram_readback_render_pass);
|
|
m_display_framebuffer = m_display_texture.CreateFramebuffer(m_display_render_pass);
|
|
if (m_vram_update_depth_framebuffer == VK_NULL_HANDLE || m_vram_readback_framebuffer == VK_NULL_HANDLE ||
|
|
m_display_framebuffer == VK_NULL_HANDLE)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
m_vram_depth_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL);
|
|
m_vram_read_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
|
|
Vulkan::DescriptorSetUpdateBuilder dsubuilder;
|
|
|
|
m_batch_descriptor_set = g_vulkan_context->AllocateGlobalDescriptorSet(m_batch_descriptor_set_layout);
|
|
m_vram_copy_descriptor_set = g_vulkan_context->AllocateGlobalDescriptorSet(m_single_sampler_descriptor_set_layout);
|
|
m_vram_read_descriptor_set = g_vulkan_context->AllocateGlobalDescriptorSet(m_single_sampler_descriptor_set_layout);
|
|
if (m_batch_descriptor_set == VK_NULL_HANDLE || m_vram_copy_descriptor_set == VK_NULL_HANDLE ||
|
|
m_vram_read_descriptor_set == VK_NULL_HANDLE)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
dsubuilder.AddBufferDescriptorWrite(m_batch_descriptor_set, 0, VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC,
|
|
m_uniform_stream_buffer.GetBuffer(), 0, sizeof(BatchUBOData));
|
|
dsubuilder.AddCombinedImageSamplerDescriptorWrite(m_batch_descriptor_set, 1, m_vram_read_texture.GetView(),
|
|
m_point_sampler, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
dsubuilder.AddCombinedImageSamplerDescriptorWrite(m_vram_copy_descriptor_set, 1, m_vram_read_texture.GetView(),
|
|
m_point_sampler, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
dsubuilder.AddCombinedImageSamplerDescriptorWrite(m_vram_read_descriptor_set, 1, m_vram_texture.GetView(),
|
|
m_point_sampler, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
dsubuilder.Update(g_vulkan_context->GetDevice());
|
|
|
|
if (old_vram_texture.IsValid())
|
|
{
|
|
const bool linear_filter = old_vram_texture.GetWidth() > m_vram_texture.GetWidth();
|
|
Log_DevPrintf("Scaling %ux%u VRAM texture to %ux%u using %s filter", old_vram_texture.GetWidth(),
|
|
old_vram_texture.GetHeight(), m_vram_texture.GetWidth(), m_vram_texture.GetHeight(),
|
|
linear_filter ? "linear" : "nearest");
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
old_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
|
|
|
|
const VkImageBlit blit{
|
|
{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{{0, 0, 0}, {static_cast<s32>(old_vram_texture.GetWidth()), static_cast<s32>(old_vram_texture.GetHeight()), 1}},
|
|
{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{{0, 0, 0}, {static_cast<s32>(m_vram_texture.GetWidth()), static_cast<s32>(m_vram_texture.GetHeight()), 1}}};
|
|
vkCmdBlitImage(cmdbuf, old_vram_texture.GetImage(), old_vram_texture.GetLayout(), m_vram_texture.GetImage(),
|
|
m_vram_texture.GetLayout(), 1, &blit, linear_filter ? VK_FILTER_LINEAR : VK_FILTER_NEAREST);
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
UpdateDepthBufferFromMaskBit();
|
|
|
|
// Can't immediately destroy because we're blitting in the current command buffer.
|
|
old_vram_texture.Destroy(true);
|
|
}
|
|
|
|
SetFullVRAMDirtyRectangle();
|
|
RestoreGraphicsAPIState();
|
|
return true;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::ClearFramebuffer()
|
|
{
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
m_vram_depth_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
|
|
static constexpr VkClearColorValue cc = {};
|
|
static constexpr VkImageSubresourceRange csrr = {VK_IMAGE_ASPECT_COLOR_BIT, 0u, 1u, 0u, 1u};
|
|
static constexpr VkClearDepthStencilValue cds = {};
|
|
static constexpr VkImageSubresourceRange dsrr = {VK_IMAGE_ASPECT_DEPTH_BIT, 0u, 1u, 0u, 1u};
|
|
vkCmdClearColorImage(cmdbuf, m_vram_texture.GetImage(), m_vram_texture.GetLayout(), &cc, 1u, &csrr);
|
|
vkCmdClearDepthStencilImage(cmdbuf, m_vram_depth_texture.GetImage(), m_vram_depth_texture.GetLayout(), &cds, 1u,
|
|
&dsrr);
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
m_vram_depth_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL);
|
|
|
|
SetFullVRAMDirtyRectangle();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::DestroyFramebuffer()
|
|
{
|
|
Vulkan::Util::SafeFreeGlobalDescriptorSet(m_batch_descriptor_set);
|
|
Vulkan::Util::SafeFreeGlobalDescriptorSet(m_vram_copy_descriptor_set);
|
|
Vulkan::Util::SafeFreeGlobalDescriptorSet(m_vram_read_descriptor_set);
|
|
|
|
Vulkan::Util::SafeDestroyFramebuffer(m_vram_framebuffer);
|
|
Vulkan::Util::SafeDestroyFramebuffer(m_vram_update_depth_framebuffer);
|
|
Vulkan::Util::SafeDestroyFramebuffer(m_vram_readback_framebuffer);
|
|
Vulkan::Util::SafeDestroyFramebuffer(m_display_framebuffer);
|
|
|
|
m_vram_read_texture.Destroy(false);
|
|
m_vram_depth_texture.Destroy(false);
|
|
m_vram_texture.Destroy(false);
|
|
m_vram_readback_texture.Destroy(false);
|
|
m_display_texture.Destroy(false);
|
|
m_vram_readback_staging_texture.Destroy(false);
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreateVertexBuffer()
|
|
{
|
|
return m_vertex_stream_buffer.Create(VK_BUFFER_USAGE_VERTEX_BUFFER_BIT, VERTEX_BUFFER_SIZE);
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreateUniformBuffer()
|
|
{
|
|
return m_uniform_stream_buffer.Create(VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT, UNIFORM_BUFFER_SIZE);
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CreateTextureBuffer()
|
|
{
|
|
if (!m_texture_stream_buffer.Create(VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT, VRAM_UPDATE_TEXTURE_BUFFER_SIZE))
|
|
return false;
|
|
|
|
Vulkan::BufferViewBuilder bvbuilder;
|
|
bvbuilder.Set(m_texture_stream_buffer.GetBuffer(), VK_FORMAT_R16_UINT, 0, m_texture_stream_buffer.GetCurrentSize());
|
|
m_texture_stream_buffer_view = bvbuilder.Create(g_vulkan_context->GetDevice());
|
|
if (m_texture_stream_buffer_view == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
m_vram_write_descriptor_set = g_vulkan_context->AllocateGlobalDescriptorSet(m_vram_write_descriptor_set_layout);
|
|
if (m_vram_write_descriptor_set == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
Vulkan::DescriptorSetUpdateBuilder dsubuilder;
|
|
dsubuilder.AddBufferViewDescriptorWrite(m_vram_write_descriptor_set, 0, VK_DESCRIPTOR_TYPE_UNIFORM_TEXEL_BUFFER,
|
|
m_texture_stream_buffer_view);
|
|
dsubuilder.Update(g_vulkan_context->GetDevice());
|
|
return true;
|
|
}
|
|
|
|
bool GPU_HW_Vulkan::CompilePipelines()
|
|
{
|
|
static constexpr std::array<VkPrimitiveTopology, 2> primitive_mapping = {
|
|
{VK_PRIMITIVE_TOPOLOGY_LINE_LIST, VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST}};
|
|
static constexpr std::array<VkPolygonMode, 2> polygon_mode_mapping = {{VK_POLYGON_MODE_LINE, VK_POLYGON_MODE_FILL}};
|
|
|
|
VkDevice device = g_vulkan_context->GetDevice();
|
|
VkPipelineCache pipeline_cache = g_vulkan_shader_cache->GetPipelineCache();
|
|
|
|
GPU_HW_ShaderGen shadergen(m_host_display->GetRenderAPI(), m_resolution_scale, m_true_color, m_scaled_dithering,
|
|
m_texture_filtering, m_supports_dual_source_blend);
|
|
|
|
// vertex shaders - [textured]
|
|
// fragment shaders - [render_mode][texture_mode][dithering][interlacing]
|
|
DimensionalArray<VkShaderModule, 2> batch_vertex_shaders{};
|
|
DimensionalArray<VkShaderModule, 2, 2, 9, 4> batch_fragment_shaders{};
|
|
Common::ScopeGuard batch_shader_guard([&batch_vertex_shaders, &batch_fragment_shaders]() {
|
|
batch_vertex_shaders.enumerate([](VkShaderModule& s) {
|
|
if (s != VK_NULL_HANDLE)
|
|
{
|
|
vkDestroyShaderModule(g_vulkan_context->GetDevice(), s, nullptr);
|
|
}
|
|
});
|
|
batch_fragment_shaders.enumerate([](VkShaderModule& s) {
|
|
if (s != VK_NULL_HANDLE)
|
|
{
|
|
vkDestroyShaderModule(g_vulkan_context->GetDevice(), s, nullptr);
|
|
}
|
|
});
|
|
});
|
|
|
|
for (u8 textured = 0; textured < 2; textured++)
|
|
{
|
|
const std::string vs = shadergen.GenerateBatchVertexShader(ConvertToBoolUnchecked(textured));
|
|
VkShaderModule shader = g_vulkan_shader_cache->GetVertexShader(vs);
|
|
if (shader == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
batch_vertex_shaders[textured] = shader;
|
|
}
|
|
|
|
for (u8 render_mode = 0; render_mode < 4; render_mode++)
|
|
{
|
|
for (u8 texture_mode = 0; texture_mode < 9; texture_mode++)
|
|
{
|
|
for (u8 dithering = 0; dithering < 2; dithering++)
|
|
{
|
|
for (u8 interlacing = 0; interlacing < 2; interlacing++)
|
|
{
|
|
const std::string fs = shadergen.GenerateBatchFragmentShader(
|
|
static_cast<BatchRenderMode>(render_mode), static_cast<TextureMode>(texture_mode),
|
|
ConvertToBoolUnchecked(dithering), ConvertToBoolUnchecked(interlacing));
|
|
|
|
VkShaderModule shader = g_vulkan_shader_cache->GetFragmentShader(fs);
|
|
if (shader == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
batch_fragment_shaders[render_mode][texture_mode][dithering][interlacing] = shader;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Vulkan::GraphicsPipelineBuilder gpbuilder;
|
|
|
|
// [primitive][depth_test][render_mode][texture_mode][transparency_mode][dithering][interlacing]
|
|
for (u8 primitive = 0; primitive < 2; primitive++)
|
|
{
|
|
for (u8 depth_test = 0; depth_test < 2; depth_test++)
|
|
{
|
|
for (u8 render_mode = 0; render_mode < 4; render_mode++)
|
|
{
|
|
for (u8 transparency_mode = 0; transparency_mode < 5; transparency_mode++)
|
|
{
|
|
for (u8 texture_mode = 0; texture_mode < 9; texture_mode++)
|
|
{
|
|
for (u8 dithering = 0; dithering < 2; dithering++)
|
|
{
|
|
for (u8 interlacing = 0; interlacing < 2; interlacing++)
|
|
{
|
|
// TODO: GS
|
|
const bool textured = (static_cast<TextureMode>(texture_mode) != TextureMode::Disabled);
|
|
|
|
gpbuilder.SetPipelineLayout(m_batch_pipeline_layout);
|
|
gpbuilder.SetRenderPass(m_vram_render_pass, 0);
|
|
|
|
gpbuilder.AddVertexBuffer(0, sizeof(BatchVertex), VK_VERTEX_INPUT_RATE_VERTEX);
|
|
gpbuilder.AddVertexAttribute(0, 0, VK_FORMAT_R32G32B32_SINT, offsetof(BatchVertex, x));
|
|
gpbuilder.AddVertexAttribute(1, 0, VK_FORMAT_R8G8B8A8_UNORM, offsetof(BatchVertex, color));
|
|
if (textured)
|
|
{
|
|
gpbuilder.AddVertexAttribute(2, 0, VK_FORMAT_R32_UINT, offsetof(BatchVertex, u));
|
|
gpbuilder.AddVertexAttribute(3, 0, VK_FORMAT_R32_UINT, offsetof(BatchVertex, texpage));
|
|
}
|
|
|
|
gpbuilder.SetPrimitiveTopology(primitive_mapping[primitive]);
|
|
gpbuilder.SetVertexShader(batch_vertex_shaders[BoolToUInt8(textured)]);
|
|
gpbuilder.SetFragmentShader(batch_fragment_shaders[render_mode][texture_mode][dithering][interlacing]);
|
|
gpbuilder.SetRasterizationState(polygon_mode_mapping[primitive], VK_CULL_MODE_NONE,
|
|
VK_FRONT_FACE_CLOCKWISE);
|
|
gpbuilder.SetDepthState(true, true,
|
|
(depth_test != 0) ? VK_COMPARE_OP_GREATER_OR_EQUAL : VK_COMPARE_OP_ALWAYS);
|
|
|
|
gpbuilder.SetNoBlendingState();
|
|
|
|
if ((static_cast<TransparencyMode>(transparency_mode) != TransparencyMode::Disabled &&
|
|
(static_cast<BatchRenderMode>(render_mode) != BatchRenderMode::TransparencyDisabled &&
|
|
static_cast<BatchRenderMode>(render_mode) != BatchRenderMode::OnlyOpaque)) ||
|
|
m_texture_filtering)
|
|
{
|
|
// TODO: Check dual-source blend support here.
|
|
gpbuilder.SetBlendAttachment(
|
|
0, true, VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_SRC1_ALPHA,
|
|
(static_cast<TransparencyMode>(transparency_mode) == TransparencyMode::BackgroundMinusForeground) ?
|
|
VK_BLEND_OP_REVERSE_SUBTRACT :
|
|
VK_BLEND_OP_ADD,
|
|
VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_ZERO, VK_BLEND_OP_ADD);
|
|
}
|
|
|
|
gpbuilder.SetDynamicViewportAndScissorState();
|
|
|
|
VkPipeline pipeline = gpbuilder.Create(device, pipeline_cache);
|
|
if (pipeline == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
m_batch_pipelines[primitive][depth_test][render_mode][texture_mode][transparency_mode][dithering]
|
|
[interlacing] = pipeline;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
batch_shader_guard.Exit();
|
|
|
|
VkShaderModule fullscreen_quad_vertex_shader =
|
|
g_vulkan_shader_cache->GetVertexShader(shadergen.GenerateScreenQuadVertexShader());
|
|
if (fullscreen_quad_vertex_shader == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
Common::ScopeGuard fullscreen_quad_vertex_shader_guard([&fullscreen_quad_vertex_shader]() {
|
|
vkDestroyShaderModule(g_vulkan_context->GetDevice(), fullscreen_quad_vertex_shader, nullptr);
|
|
});
|
|
|
|
// common state
|
|
gpbuilder.SetRenderPass(m_vram_render_pass, 0);
|
|
gpbuilder.SetPrimitiveTopology(VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST);
|
|
gpbuilder.SetNoCullRasterizationState();
|
|
gpbuilder.SetNoDepthTestState();
|
|
gpbuilder.SetNoBlendingState();
|
|
gpbuilder.SetDynamicViewportAndScissorState();
|
|
gpbuilder.SetVertexShader(fullscreen_quad_vertex_shader);
|
|
|
|
// VRAM fill
|
|
{
|
|
for (u8 interlaced = 0; interlaced < 2; interlaced++)
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(
|
|
(interlaced == 0) ? shadergen.GenerateFillFragmentShader() : shadergen.GenerateInterlacedFillFragmentShader());
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetPipelineLayout(m_no_samplers_pipeline_layout);
|
|
gpbuilder.SetFragmentShader(fs);
|
|
gpbuilder.SetDepthState(true, true, VK_COMPARE_OP_ALWAYS);
|
|
|
|
m_vram_fill_pipelines[interlaced] = gpbuilder.Create(device, pipeline_cache, false);
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
if (m_vram_fill_pipelines[interlaced] == VK_NULL_HANDLE)
|
|
return false;
|
|
}
|
|
}
|
|
|
|
// VRAM copy
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(shadergen.GenerateVRAMCopyFragmentShader());
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetPipelineLayout(m_single_sampler_pipeline_layout);
|
|
gpbuilder.SetFragmentShader(fs);
|
|
for (u8 depth_test = 0; depth_test < 2; depth_test++)
|
|
{
|
|
gpbuilder.SetDepthState((depth_test != 0), true,
|
|
(depth_test != 0) ? VK_COMPARE_OP_GREATER_OR_EQUAL : VK_COMPARE_OP_ALWAYS);
|
|
|
|
m_vram_copy_pipelines[depth_test] = gpbuilder.Create(device, pipeline_cache, false);
|
|
if (m_vram_copy_pipelines[depth_test] == VK_NULL_HANDLE)
|
|
{
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
}
|
|
|
|
// VRAM write
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(shadergen.GenerateVRAMWriteFragmentShader());
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetPipelineLayout(m_vram_write_pipeline_layout);
|
|
gpbuilder.SetFragmentShader(fs);
|
|
for (u8 depth_test = 0; depth_test < 2; depth_test++)
|
|
{
|
|
gpbuilder.SetDepthState((depth_test != 0), true,
|
|
(depth_test != 0) ? VK_COMPARE_OP_GREATER_OR_EQUAL : VK_COMPARE_OP_ALWAYS);
|
|
m_vram_write_pipelines[depth_test] = gpbuilder.Create(device, pipeline_cache, false);
|
|
if (m_vram_write_pipelines[depth_test] == VK_NULL_HANDLE)
|
|
{
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
}
|
|
|
|
// VRAM update depth
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(shadergen.GenerateVRAMUpdateDepthFragmentShader());
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetRenderPass(m_vram_update_depth_render_pass, 0);
|
|
gpbuilder.SetPipelineLayout(m_single_sampler_pipeline_layout);
|
|
gpbuilder.SetFragmentShader(fs);
|
|
gpbuilder.SetDepthState(true, true, VK_COMPARE_OP_ALWAYS);
|
|
gpbuilder.SetBlendAttachment(0, false, VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_ZERO, VK_BLEND_OP_ADD,
|
|
VK_BLEND_FACTOR_ONE, VK_BLEND_FACTOR_ZERO, VK_BLEND_OP_ADD, 0);
|
|
|
|
m_vram_update_depth_pipeline = gpbuilder.Create(device, pipeline_cache, false);
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
if (m_vram_update_depth_pipeline == VK_NULL_HANDLE)
|
|
return false;
|
|
}
|
|
|
|
gpbuilder.Clear();
|
|
|
|
// VRAM read
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(shadergen.GenerateVRAMReadFragmentShader());
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetRenderPass(m_vram_readback_render_pass, 0);
|
|
gpbuilder.SetPipelineLayout(m_single_sampler_pipeline_layout);
|
|
gpbuilder.SetVertexShader(fullscreen_quad_vertex_shader);
|
|
gpbuilder.SetFragmentShader(fs);
|
|
gpbuilder.SetNoCullRasterizationState();
|
|
gpbuilder.SetNoDepthTestState();
|
|
gpbuilder.SetNoBlendingState();
|
|
gpbuilder.SetDynamicViewportAndScissorState();
|
|
|
|
m_vram_readback_pipeline = gpbuilder.Create(device, pipeline_cache, false);
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
if (m_vram_readback_pipeline == VK_NULL_HANDLE)
|
|
return false;
|
|
}
|
|
|
|
gpbuilder.Clear();
|
|
|
|
// Display
|
|
{
|
|
gpbuilder.SetRenderPass(m_display_render_pass, 0);
|
|
gpbuilder.SetPipelineLayout(m_single_sampler_pipeline_layout);
|
|
gpbuilder.SetVertexShader(fullscreen_quad_vertex_shader);
|
|
gpbuilder.SetNoCullRasterizationState();
|
|
gpbuilder.SetNoDepthTestState();
|
|
gpbuilder.SetNoBlendingState();
|
|
gpbuilder.SetDynamicViewportAndScissorState();
|
|
|
|
for (u8 depth_24 = 0; depth_24 < 2; depth_24++)
|
|
{
|
|
for (u8 interlace_mode = 0; interlace_mode < 3; interlace_mode++)
|
|
{
|
|
VkShaderModule fs = g_vulkan_shader_cache->GetFragmentShader(shadergen.GenerateDisplayFragmentShader(
|
|
ConvertToBoolUnchecked(depth_24), static_cast<InterlacedRenderMode>(interlace_mode)));
|
|
if (fs == VK_NULL_HANDLE)
|
|
return false;
|
|
|
|
gpbuilder.SetFragmentShader(fs);
|
|
|
|
m_display_pipelines[depth_24][interlace_mode] = gpbuilder.Create(device, pipeline_cache, false);
|
|
vkDestroyShaderModule(device, fs, nullptr);
|
|
if (m_display_pipelines[depth_24][interlace_mode] == VK_NULL_HANDLE)
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
void GPU_HW_Vulkan::DestroyPipelines()
|
|
{
|
|
m_batch_pipelines.enumerate(Vulkan::Util::SafeDestroyPipeline);
|
|
|
|
for (VkPipeline& p : m_vram_fill_pipelines)
|
|
Vulkan::Util::SafeDestroyPipeline(p);
|
|
|
|
for (VkPipeline& p : m_vram_write_pipelines)
|
|
Vulkan::Util::SafeDestroyPipeline(p);
|
|
|
|
for (VkPipeline& p : m_vram_copy_pipelines)
|
|
Vulkan::Util::SafeDestroyPipeline(p);
|
|
|
|
Vulkan::Util::SafeDestroyPipeline(m_vram_readback_pipeline);
|
|
Vulkan::Util::SafeDestroyPipeline(m_vram_update_depth_pipeline);
|
|
|
|
m_display_pipelines.enumerate(Vulkan::Util::SafeDestroyPipeline);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::DrawBatchVertices(BatchRenderMode render_mode, u32 base_vertex, u32 num_vertices)
|
|
{
|
|
BeginVRAMRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
|
|
// [primitive][depth_test][render_mode][texture_mode][transparency_mode][dithering][interlacing]
|
|
VkPipeline pipeline =
|
|
m_batch_pipelines[static_cast<u8>(m_batch.primitive)][BoolToUInt8(m_batch.check_mask_before_draw)][static_cast<u8>(
|
|
render_mode)][static_cast<u8>(m_batch.texture_mode)][static_cast<u8>(m_batch.transparency_mode)]
|
|
[BoolToUInt8(m_batch.dithering)][BoolToUInt8(m_batch.interlacing)];
|
|
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, pipeline);
|
|
vkCmdDraw(cmdbuf, num_vertices, 1, base_vertex, 0);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::SetScissorFromDrawingArea()
|
|
{
|
|
int left, top, right, bottom;
|
|
CalcScissorRect(&left, &top, &right, &bottom);
|
|
|
|
Vulkan::Util::SetScissor(g_vulkan_context->GetCurrentCommandBuffer(), left, top, right - left, bottom - top);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UpdateDisplay()
|
|
{
|
|
GPU_HW::UpdateDisplay();
|
|
|
|
if (m_system->GetSettings().debugging.show_vram)
|
|
{
|
|
m_host_display->SetDisplayTexture(&m_vram_texture, m_vram_texture.GetWidth(), m_vram_texture.GetHeight(), 0, 0,
|
|
m_vram_texture.GetWidth(), m_vram_texture.GetHeight());
|
|
m_host_display->SetDisplayParameters(VRAM_WIDTH, VRAM_HEIGHT, 0, 0, VRAM_WIDTH, VRAM_HEIGHT,
|
|
static_cast<float>(VRAM_WIDTH) / static_cast<float>(VRAM_HEIGHT));
|
|
}
|
|
else
|
|
{
|
|
const u32 vram_offset_x = m_crtc_state.display_vram_left;
|
|
const u32 vram_offset_y = m_crtc_state.display_vram_top;
|
|
const u32 scaled_vram_offset_x = vram_offset_x * m_resolution_scale;
|
|
const u32 scaled_vram_offset_y = vram_offset_y * m_resolution_scale;
|
|
const u32 display_width = m_crtc_state.display_vram_width;
|
|
const u32 display_height = m_crtc_state.display_vram_height;
|
|
const u32 scaled_display_width = display_width * m_resolution_scale;
|
|
const u32 scaled_display_height = display_height * m_resolution_scale;
|
|
const InterlacedRenderMode interlaced = GetInterlacedRenderMode();
|
|
|
|
if (IsDisplayDisabled())
|
|
{
|
|
m_host_display->ClearDisplayTexture();
|
|
}
|
|
else if (!m_GPUSTAT.display_area_color_depth_24 && interlaced == InterlacedRenderMode::None &&
|
|
(scaled_vram_offset_x + scaled_display_width) <= m_vram_texture.GetWidth() &&
|
|
(scaled_vram_offset_y + scaled_display_height) <= m_vram_texture.GetHeight())
|
|
{
|
|
m_host_display->SetDisplayTexture(&m_vram_texture, m_vram_texture.GetWidth(), m_vram_texture.GetHeight(),
|
|
scaled_vram_offset_x, scaled_vram_offset_y, scaled_display_width,
|
|
scaled_display_height);
|
|
}
|
|
else
|
|
{
|
|
EndRenderPass();
|
|
|
|
const u32 reinterpret_field_offset = (interlaced != InterlacedRenderMode::None) ? GetInterlacedDisplayField() : 0;
|
|
const u32 reinterpret_start_x = m_crtc_state.regs.X * m_resolution_scale;
|
|
const u32 reinterpret_crop_left = (m_crtc_state.display_vram_left - m_crtc_state.regs.X) * m_resolution_scale;
|
|
const u32 uniforms[4] = {reinterpret_start_x, scaled_vram_offset_y + reinterpret_field_offset,
|
|
reinterpret_crop_left, reinterpret_field_offset};
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_display_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
|
|
BeginRenderPass(m_display_render_pass, m_display_framebuffer, 0, 0, scaled_display_width, scaled_display_height);
|
|
|
|
vkCmdBindPipeline(
|
|
cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
m_display_pipelines[BoolToUInt8(m_GPUSTAT.display_area_color_depth_24)][static_cast<u8>(interlaced)]);
|
|
vkCmdPushConstants(cmdbuf, m_single_sampler_pipeline_layout, VK_SHADER_STAGE_FRAGMENT_BIT, 0, sizeof(uniforms),
|
|
uniforms);
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_single_sampler_pipeline_layout, 0, 1,
|
|
&m_vram_read_descriptor_set, 0, nullptr);
|
|
Vulkan::Util::SetViewportAndScissor(cmdbuf, 0, 0, scaled_display_width, scaled_display_height);
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
|
|
EndRenderPass();
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
m_display_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
|
|
m_host_display->SetDisplayTexture(&m_display_texture, m_display_texture.GetWidth(), m_display_texture.GetHeight(),
|
|
0, 0, scaled_display_width, scaled_display_height);
|
|
|
|
RestoreGraphicsAPIState();
|
|
}
|
|
|
|
m_host_display->SetDisplayParameters(m_crtc_state.display_width, m_crtc_state.display_height,
|
|
m_crtc_state.display_origin_left, m_crtc_state.display_origin_top,
|
|
m_crtc_state.display_vram_width, m_crtc_state.display_vram_height,
|
|
m_crtc_state.display_aspect_ratio);
|
|
}
|
|
}
|
|
|
|
void GPU_HW_Vulkan::ReadVRAM(u32 x, u32 y, u32 width, u32 height)
|
|
{
|
|
// Get bounds with wrap-around handled.
|
|
const Common::Rectangle<u32> copy_rect = GetVRAMTransferBounds(x, y, width, height);
|
|
const u32 encoded_width = (copy_rect.GetWidth() + 1) / 2;
|
|
const u32 encoded_height = copy_rect.GetHeight();
|
|
|
|
EndRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
m_vram_readback_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
BeginRenderPass(m_vram_readback_render_pass, m_vram_readback_framebuffer, 0, 0, encoded_width, encoded_height);
|
|
|
|
// Encode the 24-bit texture as 16-bit.
|
|
const u32 uniforms[4] = {copy_rect.left, copy_rect.top, copy_rect.GetWidth(), copy_rect.GetHeight()};
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_vram_readback_pipeline);
|
|
vkCmdPushConstants(cmdbuf, m_single_sampler_pipeline_layout, VK_SHADER_STAGE_FRAGMENT_BIT, 0, sizeof(uniforms),
|
|
uniforms);
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_single_sampler_pipeline_layout, 0, 1,
|
|
&m_vram_read_descriptor_set, 0, nullptr);
|
|
Vulkan::Util::SetViewportAndScissor(cmdbuf, 0, 0, encoded_width, encoded_height);
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
|
|
EndRenderPass();
|
|
|
|
m_vram_readback_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
// Stage the readback.
|
|
m_vram_readback_staging_texture.CopyFromTexture(m_vram_readback_texture, 0, 0, 0, 0, 0, 0, encoded_width,
|
|
encoded_height);
|
|
|
|
// And copy it into our shadow buffer (will execute command buffer and stall).
|
|
m_vram_readback_staging_texture.ReadTexels(0, 0, encoded_width, encoded_height,
|
|
&m_vram_shadow[copy_rect.top * VRAM_WIDTH + copy_rect.left],
|
|
VRAM_WIDTH * sizeof(u16));
|
|
|
|
RestoreGraphicsAPIState();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::FillVRAM(u32 x, u32 y, u32 width, u32 height, u32 color)
|
|
{
|
|
if ((x + width) > VRAM_WIDTH || (y + height) > VRAM_HEIGHT)
|
|
{
|
|
// CPU round trip if oversized for now.
|
|
Log_WarningPrintf("Oversized VRAM fill (%u-%u, %u-%u), CPU round trip", x, x + width, y, y + height);
|
|
ReadVRAM(0, 0, VRAM_WIDTH, VRAM_HEIGHT);
|
|
GPU::FillVRAM(x, y, width, height, color);
|
|
UpdateVRAM(0, 0, VRAM_WIDTH, VRAM_HEIGHT, m_vram_shadow.data());
|
|
return;
|
|
}
|
|
|
|
GPU_HW::FillVRAM(x, y, width, height, color);
|
|
|
|
x *= m_resolution_scale;
|
|
y *= m_resolution_scale;
|
|
width *= m_resolution_scale;
|
|
height *= m_resolution_scale;
|
|
|
|
BeginVRAMRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
const VRAMFillUBOData uniforms = GetVRAMFillUBOData(x, y, width, height, color);
|
|
vkCmdPushConstants(cmdbuf, m_no_samplers_pipeline_layout, VK_SHADER_STAGE_FRAGMENT_BIT, 0, sizeof(uniforms),
|
|
&uniforms);
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
m_vram_fill_pipelines[BoolToUInt8(IsInterlacedRenderingEnabled())]);
|
|
Vulkan::Util::SetViewportAndScissor(cmdbuf, x, y, width, height);
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
|
|
RestoreGraphicsAPIState();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UpdateVRAM(u32 x, u32 y, u32 width, u32 height, const void* data)
|
|
{
|
|
if ((x + width) > VRAM_WIDTH || (y + height) > VRAM_HEIGHT)
|
|
{
|
|
// CPU round trip if oversized for now.
|
|
Log_WarningPrintf("Oversized VRAM update (%u-%u, %u-%u), CPU round trip", x, x + width, y, y + height);
|
|
ReadVRAM(0, 0, VRAM_WIDTH, VRAM_HEIGHT);
|
|
GPU::UpdateVRAM(x, y, width, height, data);
|
|
UpdateVRAM(0, 0, VRAM_WIDTH, VRAM_HEIGHT, m_vram_shadow.data());
|
|
return;
|
|
}
|
|
|
|
GPU_HW::UpdateVRAM(x, y, width, height, data);
|
|
|
|
const u32 data_size = width * height * sizeof(u16);
|
|
const u32 alignment = std::max<u32>(sizeof(u16), static_cast<u32>(g_vulkan_context->GetTexelBufferAlignment()));
|
|
if (!m_texture_stream_buffer.ReserveMemory(data_size, alignment))
|
|
{
|
|
Log_PerfPrintf("Executing command buffer while waiting for %u bytes in stream buffer", data_size);
|
|
EndRenderPass();
|
|
g_vulkan_context->ExecuteCommandBuffer(false);
|
|
RestoreGraphicsAPIState();
|
|
if (!m_texture_stream_buffer.ReserveMemory(data_size, alignment))
|
|
{
|
|
Panic("Failed to allocate space in stream buffer for VRAM write");
|
|
return;
|
|
}
|
|
}
|
|
|
|
const u32 start_index = m_texture_stream_buffer.GetCurrentOffset() / sizeof(u16);
|
|
std::memcpy(m_texture_stream_buffer.GetCurrentHostPointer(), data, data_size);
|
|
m_texture_stream_buffer.CommitMemory(data_size);
|
|
|
|
BeginVRAMRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
const VRAMWriteUBOData uniforms = {x,
|
|
y,
|
|
width,
|
|
height,
|
|
start_index,
|
|
m_GPUSTAT.set_mask_while_drawing ? 0x8000u : 0x00,
|
|
GetCurrentNormalizedVertexDepth()};
|
|
vkCmdPushConstants(cmdbuf, m_vram_write_pipeline_layout, VK_SHADER_STAGE_FRAGMENT_BIT, 0, sizeof(uniforms),
|
|
&uniforms);
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
m_vram_write_pipelines[BoolToUInt8(m_GPUSTAT.check_mask_before_draw)]);
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_vram_write_pipeline_layout, 0, 1,
|
|
&m_vram_write_descriptor_set, 0, nullptr);
|
|
|
|
// the viewport should already be set to the full vram, so just adjust the scissor
|
|
Vulkan::Util::SetScissor(cmdbuf, x * m_resolution_scale, y * m_resolution_scale, width * m_resolution_scale,
|
|
height * m_resolution_scale);
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
|
|
RestoreGraphicsAPIState();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::CopyVRAM(u32 src_x, u32 src_y, u32 dst_x, u32 dst_y, u32 width, u32 height)
|
|
{
|
|
if (UseVRAMCopyShader(src_x, src_y, dst_x, dst_y, width, height))
|
|
{
|
|
const Common::Rectangle<u32> src_bounds = GetVRAMTransferBounds(src_x, src_y, width, height);
|
|
const Common::Rectangle<u32> dst_bounds = GetVRAMTransferBounds(dst_x, dst_y, width, height);
|
|
if (m_vram_dirty_rect.Intersects(src_bounds))
|
|
UpdateVRAMReadTexture();
|
|
IncludeVRAMDityRectangle(dst_bounds);
|
|
|
|
const VRAMCopyUBOData uniforms(GetVRAMCopyUBOData(src_x, src_y, dst_x, dst_y, width, height));
|
|
const Common::Rectangle<u32> dst_bounds_scaled(dst_bounds * m_resolution_scale);
|
|
|
|
BeginVRAMRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS,
|
|
m_vram_copy_pipelines[BoolToUInt8(m_GPUSTAT.check_mask_before_draw)]);
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_single_sampler_pipeline_layout, 0, 1,
|
|
&m_vram_copy_descriptor_set, 0, nullptr);
|
|
Vulkan::Util::SetViewportAndScissor(cmdbuf, dst_bounds_scaled.left, dst_bounds_scaled.top,
|
|
dst_bounds_scaled.GetWidth(), dst_bounds_scaled.GetHeight());
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
RestoreGraphicsAPIState();
|
|
return;
|
|
}
|
|
|
|
if (m_GPUSTAT.IsMaskingEnabled())
|
|
Log_WarningPrintf("Masking enabled on VRAM copy - not implemented");
|
|
|
|
GPU_HW::CopyVRAM(src_x, src_y, dst_x, dst_y, width, height);
|
|
|
|
src_x *= m_resolution_scale;
|
|
src_y *= m_resolution_scale;
|
|
dst_x *= m_resolution_scale;
|
|
dst_y *= m_resolution_scale;
|
|
width *= m_resolution_scale;
|
|
height *= m_resolution_scale;
|
|
|
|
EndRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_GENERAL);
|
|
|
|
const VkImageCopy ic{{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{static_cast<s32>(src_x), static_cast<s32>(src_y), 0},
|
|
{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{static_cast<s32>(dst_x), static_cast<s32>(dst_y), 0},
|
|
{width, height, 1u}};
|
|
vkCmdCopyImage(cmdbuf, m_vram_texture.GetImage(), m_vram_texture.GetLayout(), m_vram_texture.GetImage(),
|
|
m_vram_texture.GetLayout(), 1, &ic);
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UpdateVRAMReadTexture()
|
|
{
|
|
EndRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
|
|
m_vram_read_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
|
|
const auto scaled_rect = m_vram_dirty_rect * m_resolution_scale;
|
|
const VkImageCopy copy{{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{static_cast<s32>(scaled_rect.left), static_cast<s32>(scaled_rect.top), 0},
|
|
{VK_IMAGE_ASPECT_COLOR_BIT, 0u, 0u, 1u},
|
|
{static_cast<s32>(scaled_rect.left), static_cast<s32>(scaled_rect.top), 0},
|
|
{scaled_rect.GetWidth(), scaled_rect.GetHeight(), 1u}};
|
|
|
|
vkCmdCopyImage(cmdbuf, m_vram_texture.GetImage(), m_vram_texture.GetLayout(), m_vram_read_texture.GetImage(),
|
|
m_vram_read_texture.GetLayout(), 1u, ©);
|
|
|
|
m_vram_read_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
GPU_HW::UpdateVRAMReadTexture();
|
|
}
|
|
|
|
void GPU_HW_Vulkan::UpdateDepthBufferFromMaskBit()
|
|
{
|
|
EndRenderPass();
|
|
|
|
VkCommandBuffer cmdbuf = g_vulkan_context->GetCurrentCommandBuffer();
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL);
|
|
|
|
BeginRenderPass(m_vram_update_depth_render_pass, m_vram_update_depth_framebuffer, 0, 0, m_vram_texture.GetWidth(),
|
|
m_vram_texture.GetHeight());
|
|
|
|
vkCmdBindPipeline(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_vram_update_depth_pipeline);
|
|
vkCmdBindDescriptorSets(cmdbuf, VK_PIPELINE_BIND_POINT_GRAPHICS, m_single_sampler_pipeline_layout, 0, 1,
|
|
&m_vram_read_descriptor_set, 0, nullptr);
|
|
Vulkan::Util::SetViewportAndScissor(cmdbuf, 0, 0, m_vram_texture.GetWidth(), m_vram_texture.GetHeight());
|
|
vkCmdDraw(cmdbuf, 3, 1, 0, 0);
|
|
|
|
EndRenderPass();
|
|
|
|
m_vram_texture.TransitionToLayout(cmdbuf, VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
|
|
RestoreGraphicsAPIState();
|
|
}
|
|
|
|
std::unique_ptr<GPU> GPU::CreateHardwareVulkanRenderer()
|
|
{
|
|
return std::make_unique<GPU_HW_Vulkan>();
|
|
}
|