627 lines
27 KiB
C++
627 lines
27 KiB
C++
// SPDX-FileCopyrightText: Copyright 2024 shadPS4 Emulator Project
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
|
|
#include <SDL_vulkan.h>
|
|
#include <core/PS4/GPU/gpu_memory.h>
|
|
#include <fmt/core.h>
|
|
#include <vulkan/vk_enum_string_helper.h>
|
|
#include <vulkan/vulkan_core.h>
|
|
#include "common/assert.h"
|
|
#include "common/debug.h"
|
|
#include "common/logging/log.h"
|
|
#include "common/singleton.h"
|
|
#include "vulkan_util.h"
|
|
|
|
#include <algorithm>
|
|
|
|
void Graphics::Vulkan::vulkanCreate(Emu::WindowCtx* ctx) {
|
|
/*Emu::VulkanExt ext;
|
|
vulkanGetInstanceExtensions(&ext);
|
|
|
|
VkApplicationInfo app_info{};
|
|
app_info.sType = VK_STRUCTURE_TYPE_APPLICATION_INFO;
|
|
app_info.pNext = nullptr;
|
|
app_info.pApplicationName = "shadps4";
|
|
app_info.applicationVersion = 1;
|
|
app_info.pEngineName = "shadps4";
|
|
app_info.engineVersion = 1;
|
|
app_info.apiVersion = VK_API_VERSION_1_2;
|
|
|
|
VkInstanceCreateInfo inst_info{};
|
|
inst_info.sType = VK_STRUCTURE_TYPE_INSTANCE_CREATE_INFO;
|
|
inst_info.pNext = nullptr;
|
|
inst_info.flags = 0;
|
|
inst_info.pApplicationInfo = &app_info;
|
|
inst_info.enabledExtensionCount = ext.required_extensions_count;
|
|
inst_info.ppEnabledExtensionNames = ext.required_extensions;
|
|
inst_info.enabledLayerCount = 0;
|
|
inst_info.ppEnabledLayerNames = nullptr;
|
|
|
|
const VkResult result = vkCreateInstance(&inst_info, nullptr, &ctx->m_graphic_ctx.m_instance);
|
|
ASSERT_MSG(result == VK_SUCCESS, "Can't create an vulkan instance");
|
|
|
|
if (SDL_Vulkan_CreateSurface(ctx->m_window, ctx->m_graphic_ctx.m_instance, NULL,
|
|
&ctx->m_surface) == SDL_FALSE) {
|
|
UNREACHABLE_MSG("Can't create an vulkan surface");
|
|
}
|
|
|
|
// TODO i am not sure if it's that it is neccesary or if it needs more
|
|
std::vector<const char*> device_extensions = {
|
|
VK_KHR_SWAPCHAIN_EXTENSION_NAME, VK_EXT_DEPTH_CLIP_ENABLE_EXTENSION_NAME,
|
|
VK_EXT_COLOR_WRITE_ENABLE_EXTENSION_NAME, "VK_KHR_maintenance1"};
|
|
|
|
Emu::VulkanQueues queues;
|
|
|
|
vulkanFindCompatiblePhysicalDevice(ctx->m_graphic_ctx.m_instance, ctx->m_surface,
|
|
device_extensions, &ctx->m_surface_capabilities,
|
|
&ctx->m_graphic_ctx.m_physical_device, &queues);
|
|
|
|
ASSERT_MSG(ctx->m_graphic_ctx.m_physical_device, "Can't find compatible vulkan device");
|
|
|
|
VkPhysicalDeviceProperties device_properties{};
|
|
vkGetPhysicalDeviceProperties(ctx->m_graphic_ctx.m_physical_device, &device_properties);
|
|
|
|
LOG_INFO(Render_Vulkan, "GFX device to be used : {}", device_properties.deviceName);
|
|
|
|
ctx->m_graphic_ctx.m_device = vulkanCreateDevice(
|
|
ctx->m_graphic_ctx.m_physical_device, ctx->m_surface, &ext, queues, device_extensions);
|
|
ASSERT_MSG(ctx->m_graphic_ctx.m_device, "Can't create vulkan device");
|
|
|
|
vulkanCreateQueues(&ctx->m_graphic_ctx, queues);
|
|
ctx->swapchain = vulkanCreateSwapchain(&ctx->m_graphic_ctx, 2);*/
|
|
}
|
|
|
|
Emu::VulkanSwapchain Graphics::Vulkan::vulkanCreateSwapchain(HLE::Libs::Graphics::GraphicCtx* ctx,
|
|
u32 image_count) {
|
|
return {};
|
|
/*auto window_ctx = Common::Singleton<Emu::WindowCtx>::Instance();
|
|
const auto& capabilities = window_ctx->m_surface_capabilities.capabilities;
|
|
Emu::VulkanSwapchain s{};
|
|
|
|
VkExtent2D extent{};
|
|
extent.width = std::clamp(ctx->screen_width, capabilities.minImageExtent.width,
|
|
capabilities.maxImageExtent.width);
|
|
extent.height = std::clamp(ctx->screen_height, capabilities.minImageExtent.height,
|
|
capabilities.maxImageExtent.height);
|
|
image_count = std::clamp(image_count, capabilities.minImageCount, capabilities.maxImageCount);
|
|
|
|
VkSwapchainCreateInfoKHR create_info{};
|
|
create_info.sType = VK_STRUCTURE_TYPE_SWAPCHAIN_CREATE_INFO_KHR;
|
|
create_info.pNext = nullptr;
|
|
create_info.flags = 0;
|
|
create_info.surface = window_ctx->m_surface;
|
|
create_info.minImageCount = image_count;
|
|
|
|
if (window_ctx->m_surface_capabilities.is_format_unorm_bgra32) {
|
|
create_info.imageFormat = VK_FORMAT_B8G8R8A8_UNORM;
|
|
create_info.imageColorSpace = VK_COLOR_SPACE_SRGB_NONLINEAR_KHR;
|
|
} else if (window_ctx->m_surface_capabilities.is_format_srgb_bgra32) {
|
|
create_info.imageFormat = VK_FORMAT_B8G8R8A8_SRGB;
|
|
create_info.imageColorSpace = VK_COLOR_SPACE_SRGB_NONLINEAR_KHR;
|
|
} else {
|
|
create_info.imageFormat = window_ctx->m_surface_capabilities.formats.at(0).format;
|
|
create_info.imageColorSpace = window_ctx->m_surface_capabilities.formats.at(0).colorSpace;
|
|
}
|
|
|
|
create_info.imageExtent = extent;
|
|
create_info.imageArrayLayers = 1;
|
|
create_info.imageUsage = VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT | VK_IMAGE_USAGE_TRANSFER_DST_BIT;
|
|
create_info.imageSharingMode = VK_SHARING_MODE_EXCLUSIVE;
|
|
create_info.queueFamilyIndexCount = 0;
|
|
create_info.pQueueFamilyIndices = nullptr;
|
|
create_info.preTransform = window_ctx->m_surface_capabilities.capabilities.currentTransform;
|
|
create_info.compositeAlpha = VK_COMPOSITE_ALPHA_OPAQUE_BIT_KHR;
|
|
create_info.presentMode = VK_PRESENT_MODE_FIFO_KHR;
|
|
create_info.clipped = VK_TRUE;
|
|
create_info.oldSwapchain = nullptr;
|
|
|
|
s.swapchain_format = create_info.imageFormat;
|
|
s.swapchain_extent = extent;
|
|
|
|
vkCreateSwapchainKHR(ctx->m_device, &create_info, nullptr, &s.swapchain);
|
|
|
|
vkGetSwapchainImagesKHR(ctx->m_device, s.swapchain, &s.swapchain_images_count, nullptr);
|
|
|
|
s.swapchain_images.resize(s.swapchain_images_count);
|
|
vkGetSwapchainImagesKHR(ctx->m_device, s.swapchain, &s.swapchain_images_count,
|
|
s.swapchain_images.data());
|
|
|
|
s.swapchain_image_views.resize(s.swapchain_images_count);
|
|
for (uint32_t i = 0; i < s.swapchain_images_count; i++) {
|
|
VkImageViewCreateInfo create_info{};
|
|
create_info.sType = VK_STRUCTURE_TYPE_IMAGE_VIEW_CREATE_INFO;
|
|
create_info.pNext = nullptr;
|
|
create_info.flags = 0;
|
|
create_info.image = s.swapchain_images[i];
|
|
create_info.viewType = VK_IMAGE_VIEW_TYPE_2D;
|
|
create_info.format = s.swapchain_format;
|
|
create_info.components.r = VK_COMPONENT_SWIZZLE_IDENTITY;
|
|
create_info.components.g = VK_COMPONENT_SWIZZLE_IDENTITY;
|
|
create_info.components.b = VK_COMPONENT_SWIZZLE_IDENTITY;
|
|
create_info.components.a = VK_COMPONENT_SWIZZLE_IDENTITY;
|
|
create_info.subresourceRange.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
|
|
create_info.subresourceRange.baseArrayLayer = 0;
|
|
create_info.subresourceRange.baseMipLevel = 0;
|
|
create_info.subresourceRange.layerCount = 1;
|
|
create_info.subresourceRange.levelCount = 1;
|
|
|
|
vkCreateImageView(ctx->m_device, &create_info, nullptr, &s.swapchain_image_views[i]);
|
|
}
|
|
|
|
ASSERT_MSG(s.swapchain, "Could not create swapchain");
|
|
s.current_index = static_cast<uint32_t>(-1);
|
|
|
|
VkSemaphoreCreateInfo present_complete_info{};
|
|
present_complete_info.sType = VK_STRUCTURE_TYPE_SEMAPHORE_CREATE_INFO;
|
|
present_complete_info.pNext = nullptr;
|
|
present_complete_info.flags = 0;
|
|
|
|
auto result = vkCreateSemaphore(ctx->m_device, &present_complete_info, nullptr,
|
|
&s.present_complete_semaphore);
|
|
|
|
VkFenceCreateInfo fence_info{};
|
|
fence_info.sType = VK_STRUCTURE_TYPE_FENCE_CREATE_INFO;
|
|
fence_info.pNext = nullptr;
|
|
fence_info.flags = 0;
|
|
|
|
result = vkCreateFence(ctx->m_device, &fence_info, nullptr, &s.present_complete_fence);
|
|
ASSERT_MSG(result == VK_SUCCESS, "Can't create vulkan fence");
|
|
|
|
return s;*/
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanCreateQueues(HLE::Libs::Graphics::GraphicCtx* ctx,
|
|
const Emu::VulkanQueues& queues) {
|
|
auto get_queue = [ctx](int id, const Emu::VulkanQueueInfo& info, bool with_mutex = false) {
|
|
ctx->queues[id].family = info.family;
|
|
ctx->queues[id].index = info.index;
|
|
vkGetDeviceQueue(ctx->m_device, ctx->queues[id].family, ctx->queues[id].index,
|
|
&ctx->queues[id].vk_queue);
|
|
if (with_mutex) {
|
|
ctx->queues[id].mutex = std::make_unique<std::mutex>();
|
|
}
|
|
};
|
|
|
|
get_queue(VULKAN_QUEUE_GFX, queues.graphics.at(0));
|
|
get_queue(VULKAN_QUEUE_UTIL, queues.transfer.at(0));
|
|
get_queue(VULKAN_QUEUE_PRESENT, queues.present.at(0));
|
|
|
|
for (int id = 0; id < VULKAN_QUEUE_COMPUTE_NUM; id++) {
|
|
bool with_mutex = (VULKAN_QUEUE_COMPUTE_NUM == queues.compute.size());
|
|
get_queue(id, queues.compute.at(id % queues.compute.size()), with_mutex);
|
|
}
|
|
}
|
|
|
|
VkDevice Graphics::Vulkan::vulkanCreateDevice(VkPhysicalDevice physical_device,
|
|
VkSurfaceKHR surface, const Emu::VulkanExt* r,
|
|
const Emu::VulkanQueues& queues,
|
|
const std::vector<const char*>& device_extensions) {
|
|
/*std::vector<VkDeviceQueueCreateInfo> queue_create_info(queues.family_count);
|
|
std::vector<std::vector<float>> queue_priority(queues.family_count);
|
|
uint32_t queue_create_info_num = 0;
|
|
|
|
for (uint32_t i = 0; i < queues.family_count; i++) {
|
|
if (queues.family_used[i] != 0) {
|
|
for (uint32_t pi = 0; pi < queues.family_used[i]; pi++) {
|
|
queue_priority[queue_create_info_num].push_back(1.0f);
|
|
}
|
|
|
|
queue_create_info[queue_create_info_num].sType =
|
|
VK_STRUCTURE_TYPE_DEVICE_QUEUE_CREATE_INFO;
|
|
queue_create_info[queue_create_info_num].pNext = nullptr;
|
|
queue_create_info[queue_create_info_num].flags = 0;
|
|
queue_create_info[queue_create_info_num].queueFamilyIndex = i;
|
|
queue_create_info[queue_create_info_num].queueCount = queues.family_used[i];
|
|
queue_create_info[queue_create_info_num].pQueuePriorities =
|
|
queue_priority[queue_create_info_num].data();
|
|
|
|
queue_create_info_num++;
|
|
}
|
|
}
|
|
|
|
VkPhysicalDeviceFeatures device_features{};
|
|
// TODO add neccesary device features
|
|
|
|
VkDeviceCreateInfo create_info{};
|
|
create_info.sType = VK_STRUCTURE_TYPE_DEVICE_CREATE_INFO;
|
|
create_info.pNext = nullptr;
|
|
create_info.flags = 0;
|
|
create_info.pQueueCreateInfos = queue_create_info.data();
|
|
create_info.queueCreateInfoCount = queue_create_info_num;
|
|
create_info.enabledLayerCount = 0;
|
|
create_info.ppEnabledLayerNames = nullptr;
|
|
create_info.enabledExtensionCount = device_extensions.size();
|
|
create_info.ppEnabledExtensionNames = device_extensions.data();
|
|
create_info.pEnabledFeatures = &device_features;
|
|
|
|
VkDevice device = nullptr;
|
|
|
|
vkCreateDevice(physical_device, &create_info, nullptr, &device);
|
|
|
|
return device;*/
|
|
}
|
|
void Graphics::Vulkan::vulkanGetInstanceExtensions(Emu::VulkanExt* ext) {
|
|
/*u32 required_extensions_count = 0;
|
|
u32 available_extensions_count = 0;
|
|
u32 available_layers_count = 0;
|
|
ext->required_extensions = SDL_Vulkan_GetInstanceExtensions(&required_extensions_count);
|
|
ext->required_extensions_count = required_extensions_count;
|
|
|
|
vkEnumerateInstanceExtensionProperties(nullptr, &available_extensions_count, nullptr);
|
|
|
|
ext->available_extensions = std::vector<VkExtensionProperties>(available_extensions_count);
|
|
|
|
vkEnumerateInstanceExtensionProperties(nullptr, &available_extensions_count,
|
|
ext->available_extensions.data());
|
|
|
|
vkEnumerateInstanceLayerProperties(&available_layers_count, nullptr);
|
|
ext->available_layers = std::vector<VkLayerProperties>(available_layers_count);
|
|
vkEnumerateInstanceLayerProperties(&available_layers_count, ext->available_layers.data());
|
|
|
|
// for (const char* ext : ext->required_extensions) {
|
|
// LOG_INFO(Render_Vulkan, "Vulkan required extension = {}", ext);
|
|
//}
|
|
|
|
for (const auto& ext : ext->available_extensions) {
|
|
LOG_INFO(Render_Vulkan, "Vulkan available extension: {}, version = {}", ext.extensionName,
|
|
ext.specVersion);
|
|
}
|
|
|
|
for (const auto& l : ext->available_layers) {
|
|
LOG_INFO(Render_Vulkan,
|
|
"Vulkan available layer: {}, specVersion = {}, implVersion = {}, {}", l.layerName,
|
|
l.specVersion, l.implementationVersion, l.description);
|
|
}*/
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanFindCompatiblePhysicalDevice(
|
|
VkInstance instance, VkSurfaceKHR surface, const std::vector<const char*>& device_extensions,
|
|
Emu::VulkanSurfaceCapabilities* out_capabilities, VkPhysicalDevice* out_device,
|
|
Emu::VulkanQueues* out_queues) {
|
|
/*u32 count_devices = 0;
|
|
vkEnumeratePhysicalDevices(instance, &count_devices, nullptr);
|
|
|
|
std::vector<VkPhysicalDevice> devices(count_devices);
|
|
vkEnumeratePhysicalDevices(instance, &count_devices, devices.data());
|
|
|
|
VkPhysicalDevice found_best_device = nullptr;
|
|
Emu::VulkanQueues found_best_queues;
|
|
|
|
for (const auto& device : devices) {
|
|
VkPhysicalDeviceProperties device_properties{};
|
|
VkPhysicalDeviceFeatures2 device_features2{};
|
|
|
|
vkGetPhysicalDeviceProperties(device, &device_properties);
|
|
vkGetPhysicalDeviceFeatures2(device, &device_features2);
|
|
if (device_properties.deviceType == VK_PHYSICAL_DEVICE_TYPE_INTEGRATED_GPU) {
|
|
continue; // we don't want integrated gpu for now .Later we will check the requirements
|
|
// and see what we can support (TODO fix me)
|
|
}
|
|
LOG_INFO(Render_Vulkan, "Vulkan device: {}", device_properties.deviceName);
|
|
|
|
auto qs = vulkanFindQueues(device, surface);
|
|
|
|
vulkanGetSurfaceCapabilities(device, surface, out_capabilities);
|
|
|
|
found_best_device = device;
|
|
found_best_queues = qs;
|
|
}
|
|
*out_device = found_best_device;
|
|
*out_queues = found_best_queues;*/
|
|
}
|
|
|
|
Emu::VulkanQueues Graphics::Vulkan::vulkanFindQueues(VkPhysicalDevice device,
|
|
VkSurfaceKHR surface) {
|
|
Emu::VulkanQueues qs;
|
|
|
|
/*u32 queue_family_count = 0;
|
|
vkGetPhysicalDeviceQueueFamilyProperties(device, &queue_family_count, nullptr);
|
|
std::vector<VkQueueFamilyProperties> queue_families(queue_family_count);
|
|
vkGetPhysicalDeviceQueueFamilyProperties(device, &queue_family_count, queue_families.data());
|
|
|
|
qs.family_count = queue_family_count;
|
|
|
|
u32 family = 0;
|
|
for (auto& f : queue_families) {
|
|
VkBool32 presentation_supported = VK_FALSE;
|
|
vkGetPhysicalDeviceSurfaceSupportKHR(device, family, surface, &presentation_supported);
|
|
|
|
LOG_INFO(Render_Vulkan, "queue family: {}, count = {}, present = {}",
|
|
string_VkQueueFlags(f.queueFlags).c_str(), f.queueCount,
|
|
(presentation_supported == VK_TRUE ? "true" : "false"));
|
|
for (uint32_t i = 0; i < f.queueCount; i++) {
|
|
Emu::VulkanQueueInfo info;
|
|
info.family = family;
|
|
info.index = i;
|
|
info.is_graphics = (f.queueFlags & VK_QUEUE_GRAPHICS_BIT) != 0;
|
|
info.is_compute = (f.queueFlags & VK_QUEUE_COMPUTE_BIT) != 0;
|
|
info.is_transfer = (f.queueFlags & VK_QUEUE_TRANSFER_BIT) != 0;
|
|
info.is_present = (presentation_supported == VK_TRUE);
|
|
|
|
qs.available.push_back(info);
|
|
}
|
|
|
|
qs.family_used.push_back(0);
|
|
|
|
family++;
|
|
}
|
|
u32 index = 0;
|
|
for (u32 i = 0; i < VULKAN_QUEUE_GRAPHICS_NUM; i++) {
|
|
for (const auto& idx : qs.available) {
|
|
if (idx.is_graphics) {
|
|
qs.family_used[qs.available.at(index).family]++;
|
|
qs.graphics.push_back(qs.available.at(index));
|
|
qs.available.erase(qs.available.begin() + index);
|
|
break;
|
|
}
|
|
index++;
|
|
}
|
|
}
|
|
index = 0;
|
|
for (u32 i = 0; i < VULKAN_QUEUE_COMPUTE_NUM; i++) {
|
|
for (const auto& idx : qs.available) {
|
|
if (idx.is_graphics) {
|
|
qs.family_used[qs.available.at(index).family]++;
|
|
qs.compute.push_back(qs.available.at(index));
|
|
qs.available.erase(qs.available.begin() + index);
|
|
break;
|
|
}
|
|
index++;
|
|
}
|
|
}
|
|
index = 0;
|
|
for (uint32_t i = 0; i < VULKAN_QUEUE_TRANSFER_NUM; i++) {
|
|
for (const auto& idx : qs.available) {
|
|
if (idx.is_graphics) {
|
|
qs.family_used[qs.available.at(index).family]++;
|
|
qs.transfer.push_back(qs.available.at(index));
|
|
qs.available.erase(qs.available.begin() + index);
|
|
break;
|
|
}
|
|
index++;
|
|
}
|
|
}
|
|
index = 0;
|
|
for (uint32_t i = 0; i < VULKAN_QUEUE_PRESENT_NUM; i++) {
|
|
for (const auto& idx : qs.available) {
|
|
if (idx.is_graphics) {
|
|
qs.family_used[qs.available.at(index).family]++;
|
|
qs.present.push_back(qs.available.at(index));
|
|
qs.available.erase(qs.available.begin() + index);
|
|
break;
|
|
}
|
|
index++;
|
|
}
|
|
}*/
|
|
return qs;
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanGetSurfaceCapabilities(VkPhysicalDevice physical_device,
|
|
VkSurfaceKHR surface,
|
|
Emu::VulkanSurfaceCapabilities* surfaceCap) {
|
|
/*vkGetPhysicalDeviceSurfaceCapabilitiesKHR(physical_device, surface,
|
|
&surfaceCap->capabilities);
|
|
|
|
uint32_t formats_count = 0;
|
|
vkGetPhysicalDeviceSurfaceFormatsKHR(physical_device, surface, &formats_count, nullptr);
|
|
|
|
surfaceCap->formats = std::vector<VkSurfaceFormatKHR>(formats_count);
|
|
vkGetPhysicalDeviceSurfaceFormatsKHR(physical_device, surface, &formats_count,
|
|
surfaceCap->formats.data());
|
|
|
|
uint32_t present_modes_count = 0;
|
|
vkGetPhysicalDeviceSurfacePresentModesKHR(physical_device, surface, &present_modes_count,
|
|
nullptr);
|
|
|
|
surfaceCap->present_modes = std::vector<VkPresentModeKHR>(present_modes_count);
|
|
vkGetPhysicalDeviceSurfacePresentModesKHR(physical_device, surface, &present_modes_count,
|
|
surfaceCap->present_modes.data());
|
|
|
|
for (const auto& f : surfaceCap->formats) {
|
|
if (f.format == VK_FORMAT_B8G8R8A8_SRGB &&
|
|
f.colorSpace == VK_COLOR_SPACE_SRGB_NONLINEAR_KHR) {
|
|
surfaceCap->is_format_srgb_bgra32 = true;
|
|
break;
|
|
}
|
|
if (f.format == VK_FORMAT_B8G8R8A8_UNORM &&
|
|
f.colorSpace == VK_COLOR_SPACE_SRGB_NONLINEAR_KHR) {
|
|
surfaceCap->is_format_unorm_bgra32 = true;
|
|
break;
|
|
}
|
|
}*/
|
|
}
|
|
|
|
static void set_image_layout(VkCommandBuffer buffer, HLE::Libs::Graphics::VulkanImage* dst_image,
|
|
uint32_t base_level, uint32_t levels, VkImageAspectFlags aspect_mask,
|
|
VkImageLayout old_image_layout, VkImageLayout new_image_layout) {
|
|
VkImageMemoryBarrier imageMemoryBarrier{};
|
|
imageMemoryBarrier.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER;
|
|
imageMemoryBarrier.pNext = nullptr;
|
|
imageMemoryBarrier.srcAccessMask = 0;
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
imageMemoryBarrier.oldLayout = old_image_layout;
|
|
imageMemoryBarrier.newLayout = new_image_layout;
|
|
imageMemoryBarrier.srcQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
|
|
imageMemoryBarrier.dstQueueFamilyIndex = VK_QUEUE_FAMILY_IGNORED;
|
|
imageMemoryBarrier.image = dst_image->image;
|
|
imageMemoryBarrier.subresourceRange.aspectMask = aspect_mask;
|
|
imageMemoryBarrier.subresourceRange.baseMipLevel = base_level;
|
|
imageMemoryBarrier.subresourceRange.levelCount = levels;
|
|
imageMemoryBarrier.subresourceRange.baseArrayLayer = 0;
|
|
imageMemoryBarrier.subresourceRange.layerCount = 1;
|
|
|
|
if (old_image_layout == VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL) {
|
|
imageMemoryBarrier.srcAccessMask = 0;
|
|
}
|
|
|
|
if (new_image_layout == VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL) {
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
}
|
|
|
|
if (new_image_layout == VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL) {
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
}
|
|
|
|
if (old_image_layout == VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL) {
|
|
imageMemoryBarrier.srcAccessMask = 0;
|
|
}
|
|
|
|
if (old_image_layout == VK_IMAGE_LAYOUT_PREINITIALIZED) {
|
|
imageMemoryBarrier.srcAccessMask = VK_ACCESS_HOST_WRITE_BIT;
|
|
}
|
|
|
|
if (new_image_layout == VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL) {
|
|
imageMemoryBarrier.srcAccessMask = 0;
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
}
|
|
|
|
if (new_image_layout == VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL) {
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
}
|
|
|
|
if (new_image_layout == VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL) {
|
|
imageMemoryBarrier.dstAccessMask = 0;
|
|
}
|
|
|
|
VkPipelineStageFlags src_stages = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
|
|
VkPipelineStageFlags dest_stages = VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT;
|
|
|
|
// vkCmdPipelineBarrier(buffer, src_stages, dest_stages, 0, 0, nullptr, 0, nullptr, 1,
|
|
// &imageMemoryBarrier);
|
|
|
|
dst_image->layout = new_image_layout;
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanBlitImage(GPU::CommandBuffer* buffer,
|
|
HLE::Libs::Graphics::VulkanImage* src_image,
|
|
Emu::VulkanSwapchain* dst_swapchain) {
|
|
auto* vk_buffer = buffer->getPool()->buffers[buffer->getIndex()];
|
|
|
|
HLE::Libs::Graphics::VulkanImage swapchain_image(HLE::Libs::Graphics::VulkanImageType::Unknown);
|
|
|
|
swapchain_image.image = dst_swapchain->swapchain_images[dst_swapchain->current_index];
|
|
swapchain_image.layout = VK_IMAGE_LAYOUT_UNDEFINED;
|
|
|
|
set_image_layout(vk_buffer, src_image, 0, 1, VK_IMAGE_ASPECT_COLOR_BIT,
|
|
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL,
|
|
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL);
|
|
set_image_layout(vk_buffer, &swapchain_image, 0, 1, VK_IMAGE_ASPECT_COLOR_BIT,
|
|
VK_IMAGE_LAYOUT_UNDEFINED, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
|
|
VkImageBlit region{};
|
|
region.srcSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
|
|
region.srcSubresource.mipLevel = 0;
|
|
region.srcSubresource.baseArrayLayer = 0;
|
|
region.srcSubresource.layerCount = 1;
|
|
region.srcOffsets[0].x = 0;
|
|
region.srcOffsets[0].y = 0;
|
|
region.srcOffsets[0].z = 0;
|
|
region.srcOffsets[1].x = static_cast<int>(src_image->extent.width);
|
|
region.srcOffsets[1].y = static_cast<int>(src_image->extent.height);
|
|
region.srcOffsets[1].z = 1;
|
|
region.dstSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
|
|
region.dstSubresource.mipLevel = 0;
|
|
region.dstSubresource.baseArrayLayer = 0;
|
|
region.dstSubresource.layerCount = 1;
|
|
region.dstOffsets[0].x = 0;
|
|
region.dstOffsets[0].y = 0;
|
|
region.dstOffsets[0].z = 0;
|
|
region.dstOffsets[1].x = static_cast<int>(dst_swapchain->swapchain_extent.width);
|
|
region.dstOffsets[1].y = static_cast<int>(dst_swapchain->swapchain_extent.height);
|
|
region.dstOffsets[1].z = 1;
|
|
|
|
// vkCmdBlitImage(vk_buffer, src_image->image, VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
|
|
// swapchain_image.image, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1, ®ion,
|
|
// VK_FILTER_LINEAR);
|
|
|
|
set_image_layout(vk_buffer, src_image, 0, 1, VK_IMAGE_ASPECT_COLOR_BIT,
|
|
VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
|
|
VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL);
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanFillImage(HLE::Libs::Graphics::GraphicCtx* ctx,
|
|
HLE::Libs::Graphics::VulkanImage* dst_image,
|
|
const void* src_data, u64 size, u32 src_pitch,
|
|
u64 dst_layout) {
|
|
HLE::Libs::Graphics::VulkanBuffer staging_buffer{};
|
|
staging_buffer.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT;
|
|
staging_buffer.memory.property =
|
|
VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
|
|
vulkanCreateBuffer(ctx, size, &staging_buffer);
|
|
|
|
void* data = nullptr;
|
|
// vkMapMemory(ctx->m_device, staging_buffer.memory.memory, staging_buffer.memory.offset,
|
|
// staging_buffer.memory.requirements.size, 0, &data);
|
|
// std::memcpy(data, src_data, size);
|
|
// vkUnmapMemory(ctx->m_device, staging_buffer.memory.memory);
|
|
|
|
GPU::CommandBuffer buffer(9);
|
|
|
|
buffer.begin();
|
|
vulkanBufferToImage(&buffer, &staging_buffer, src_pitch, dst_image, dst_layout);
|
|
buffer.end();
|
|
buffer.execute();
|
|
buffer.waitForFence();
|
|
|
|
vulkanDeleteBuffer(ctx, &staging_buffer);
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanBufferToImage(GPU::CommandBuffer* buffer,
|
|
HLE::Libs::Graphics::VulkanBuffer* src_buffer,
|
|
u32 src_pitch,
|
|
HLE::Libs::Graphics::VulkanImage* dst_image,
|
|
u64 dst_layout) {
|
|
auto* vk_buffer = buffer->getPool()->buffers[buffer->getIndex()];
|
|
|
|
set_image_layout(vk_buffer, dst_image, 0, 1, VK_IMAGE_ASPECT_COLOR_BIT,
|
|
VK_IMAGE_LAYOUT_UNDEFINED, VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL);
|
|
|
|
VkBufferImageCopy region{};
|
|
region.bufferOffset = 0;
|
|
region.bufferRowLength = (src_pitch != dst_image->extent.width ? src_pitch : 0);
|
|
region.bufferImageHeight = 0;
|
|
|
|
region.imageSubresource.aspectMask = VK_IMAGE_ASPECT_COLOR_BIT;
|
|
region.imageSubresource.mipLevel = 0;
|
|
region.imageSubresource.baseArrayLayer = 0;
|
|
region.imageSubresource.layerCount = 1;
|
|
|
|
region.imageOffset = {0, 0, 0};
|
|
region.imageExtent = {dst_image->extent.width, dst_image->extent.height, 1};
|
|
|
|
// vkCmdCopyBufferToImage(vk_buffer, src_buffer->buffer, dst_image->image,
|
|
// VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, 1, ®ion);
|
|
|
|
set_image_layout(vk_buffer, dst_image, 0, 1, VK_IMAGE_ASPECT_COLOR_BIT,
|
|
VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL, static_cast<VkImageLayout>(dst_layout));
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanCreateBuffer(HLE::Libs::Graphics::GraphicCtx* ctx, u64 size,
|
|
HLE::Libs::Graphics::VulkanBuffer* buffer) {
|
|
VkBufferCreateInfo buffer_info{};
|
|
buffer_info.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
|
|
buffer_info.size = size;
|
|
buffer_info.usage = buffer->usage;
|
|
buffer_info.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
|
|
|
|
// vkCreateBuffer(ctx->m_device, &buffer_info, nullptr, &buffer->buffer);
|
|
|
|
// vkGetBufferMemoryRequirements(ctx->m_device, buffer->buffer, &buffer->memory.requirements);
|
|
|
|
bool allocated = GPU::vulkanAllocateMemory(ctx, &buffer->memory);
|
|
if (!allocated) {
|
|
fmt::print("Can't allocate vulkan\n");
|
|
std::exit(0);
|
|
}
|
|
// vkBindBufferMemory(ctx->m_device, buffer->buffer, buffer->memory.memory,
|
|
// buffer->memory.offset);
|
|
}
|
|
|
|
void Graphics::Vulkan::vulkanDeleteBuffer(HLE::Libs::Graphics::GraphicCtx* ctx,
|
|
HLE::Libs::Graphics::VulkanBuffer* buffer) {
|
|
// vkDestroyBuffer(ctx->m_device, buffer->buffer, nullptr);
|
|
// vkFreeMemory(ctx->m_device, buffer->memory.memory, nullptr);
|
|
buffer->memory.memory = nullptr;
|
|
buffer->buffer = nullptr;
|
|
}
|