/* Wine Vulkan ICD implementation * * Copyright 2017 Roderick Colenbrander * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA */ #include "config.h" #include #include #include "windef.h" #include "winbase.h" #include "winreg.h" #include "winuser.h" #include "initguid.h" #include "devguid.h" #include "setupapi.h" #include "vulkan_private.h" WINE_DEFAULT_DEBUG_CHANNEL(vulkan); DEFINE_DEVPROPKEY(DEVPROPKEY_GPU_LUID, 0x60b193cb, 0x5276, 0x4d0f, 0x96, 0xfc, 0xf1, 0x73, 0xab, 0xad, 0x3e, 0xc6, 2); DEFINE_DEVPROPKEY(WINE_DEVPROPKEY_GPU_VULKAN_UUID, 0x233a9ef3, 0xafc4, 0x4abd, 0xb5, 0x64, 0xc3, 0x2f, 0x21, 0xf1, 0x53, 0x5c, 2); /* For now default to 4 as it felt like a reasonable version feature wise to support. * Don't support the optional vk_icdGetPhysicalDeviceProcAddr introduced in this version * as it is unlikely we will implement physical device extensions, which the loader is not * aware of. Version 5 adds more extensive version checks. Something to tackle later. */ #define WINE_VULKAN_ICD_VERSION 4 #define wine_vk_find_struct(s, t) wine_vk_find_struct_((void *)s, VK_STRUCTURE_TYPE_##t) static void *wine_vk_find_struct_(void *s, VkStructureType t) { VkBaseOutStructure *header; for (header = s; header; header = header->pNext) { if (header->sType == t) return header; } return NULL; } static void *wine_vk_get_global_proc_addr(const char *name); static HINSTANCE hinstance; static const struct vulkan_funcs *vk_funcs; static VkResult (*p_vkEnumerateInstanceVersion)(uint32_t *version); void WINAPI wine_vkGetPhysicalDeviceProperties(VkPhysicalDevice physical_device, VkPhysicalDeviceProperties *properties); static void wine_vk_physical_device_free(struct VkPhysicalDevice_T *phys_dev) { if (!phys_dev) return; heap_free(phys_dev->extensions); heap_free(phys_dev); } static struct VkPhysicalDevice_T *wine_vk_physical_device_alloc(struct VkInstance_T *instance, VkPhysicalDevice phys_dev) { struct VkPhysicalDevice_T *object; uint32_t num_host_properties, num_properties = 0; VkExtensionProperties *host_properties = NULL; VkResult res; unsigned int i, j; if (!(object = heap_alloc_zero(sizeof(*object)))) return NULL; object->base.loader_magic = VULKAN_ICD_MAGIC_VALUE; object->instance = instance; object->phys_dev = phys_dev; res = instance->funcs.p_vkEnumerateDeviceExtensionProperties(phys_dev, NULL, &num_host_properties, NULL); if (res != VK_SUCCESS) { ERR("Failed to enumerate device extensions, res=%d\n", res); goto err; } host_properties = heap_calloc(num_host_properties, sizeof(*host_properties)); if (!host_properties) { ERR("Failed to allocate memory for device properties!\n"); goto err; } res = instance->funcs.p_vkEnumerateDeviceExtensionProperties(phys_dev, NULL, &num_host_properties, host_properties); if (res != VK_SUCCESS) { ERR("Failed to enumerate device extensions, res=%d\n", res); goto err; } /* Count list of extensions for which we have an implementation. * TODO: perform translation for platform specific extensions. */ for (i = 0; i < num_host_properties; i++) { if (wine_vk_device_extension_supported(host_properties[i].extensionName)) { TRACE("Enabling extension '%s' for physical device %p\n", host_properties[i].extensionName, object); num_properties++; } else { TRACE("Skipping extension '%s', no implementation found in winevulkan.\n", host_properties[i].extensionName); } } TRACE("Host supported extensions %u, Wine supported extensions %u\n", num_host_properties, num_properties); if (!(object->extensions = heap_calloc(num_properties, sizeof(*object->extensions)))) { ERR("Failed to allocate memory for device extensions!\n"); goto err; } for (i = 0, j = 0; i < num_host_properties; i++) { if (wine_vk_device_extension_supported(host_properties[i].extensionName)) { object->extensions[j] = host_properties[i]; j++; } } object->extension_count = num_properties; heap_free(host_properties); return object; err: wine_vk_physical_device_free(object); heap_free(host_properties); return NULL; } static void wine_vk_free_command_buffers(struct VkDevice_T *device, struct wine_cmd_pool *pool, uint32_t count, const VkCommandBuffer *buffers) { unsigned int i; for (i = 0; i < count; i++) { if (!buffers[i]) continue; device->funcs.p_vkFreeCommandBuffers(device->device, pool->command_pool, 1, &buffers[i]->command_buffer); list_remove(&buffers[i]->pool_link); heap_free(buffers[i]); } } static struct VkQueue_T *wine_vk_device_alloc_queues(struct VkDevice_T *device, uint32_t family_index, uint32_t queue_count, VkDeviceQueueCreateFlags flags) { VkDeviceQueueInfo2 queue_info; struct VkQueue_T *queues; unsigned int i; if (!(queues = heap_calloc(queue_count, sizeof(*queues)))) { ERR("Failed to allocate memory for queues\n"); return NULL; } for (i = 0; i < queue_count; i++) { struct VkQueue_T *queue = &queues[i]; queue->base.loader_magic = VULKAN_ICD_MAGIC_VALUE; queue->device = device; queue->flags = flags; /* The Vulkan spec says: * * "vkGetDeviceQueue must only be used to get queues that were created * with the flags parameter of VkDeviceQueueCreateInfo set to zero." */ if (flags && device->funcs.p_vkGetDeviceQueue2) { queue_info.sType = VK_STRUCTURE_TYPE_DEVICE_QUEUE_INFO_2; queue_info.pNext = NULL; queue_info.flags = flags; queue_info.queueFamilyIndex = family_index; queue_info.queueIndex = i; device->funcs.p_vkGetDeviceQueue2(device->device, &queue_info, &queue->queue); } else { device->funcs.p_vkGetDeviceQueue(device->device, family_index, i, &queue->queue); } } return queues; } static void wine_vk_device_free_create_info(VkDeviceCreateInfo *create_info) { VkDeviceGroupDeviceCreateInfo *group_info; if ((group_info = wine_vk_find_struct(create_info, DEVICE_GROUP_DEVICE_CREATE_INFO))) { heap_free((void *)group_info->pPhysicalDevices); } free_VkDeviceCreateInfo_struct_chain(create_info); } static VkResult wine_vk_device_convert_create_info(const VkDeviceCreateInfo *src, VkDeviceCreateInfo *dst) { VkDeviceGroupDeviceCreateInfo *group_info; unsigned int i; VkResult res; *dst = *src; if ((res = convert_VkDeviceCreateInfo_struct_chain(src->pNext, dst)) < 0) { WARN("Failed to convert VkDeviceCreateInfo pNext chain, res=%d.\n", res); return res; } /* FIXME: convert_VkDeviceCreateInfo_struct_chain() should unwrap handles for us. */ if ((group_info = wine_vk_find_struct(dst, DEVICE_GROUP_DEVICE_CREATE_INFO))) { VkPhysicalDevice *physical_devices; if (!(physical_devices = heap_calloc(group_info->physicalDeviceCount, sizeof(*physical_devices)))) { free_VkDeviceCreateInfo_struct_chain(dst); return VK_ERROR_OUT_OF_HOST_MEMORY; } for (i = 0; i < group_info->physicalDeviceCount; ++i) { physical_devices[i] = group_info->pPhysicalDevices[i]->phys_dev; } group_info->pPhysicalDevices = physical_devices; } /* Should be filtered out by loader as ICDs don't support layers. */ dst->enabledLayerCount = 0; dst->ppEnabledLayerNames = NULL; TRACE("Enabled %u extensions.\n", dst->enabledExtensionCount); for (i = 0; i < dst->enabledExtensionCount; i++) { const char *extension_name = dst->ppEnabledExtensionNames[i]; TRACE("Extension %u: %s.\n", i, debugstr_a(extension_name)); if (!wine_vk_device_extension_supported(extension_name)) { WARN("Extension %s is not supported.\n", debugstr_a(extension_name)); wine_vk_device_free_create_info(dst); return VK_ERROR_EXTENSION_NOT_PRESENT; } } return VK_SUCCESS; } /* Helper function used for freeing a device structure. This function supports full * and partial object cleanups and can thus be used for vkCreateDevice failures. */ static void wine_vk_device_free(struct VkDevice_T *device) { if (!device) return; if (device->queues) { unsigned int i; for (i = 0; i < device->max_queue_families; i++) { heap_free(device->queues[i]); } heap_free(device->queues); device->queues = NULL; } if (device->device && device->funcs.p_vkDestroyDevice) { device->funcs.p_vkDestroyDevice(device->device, NULL /* pAllocator */); } heap_free(device); } static BOOL WINAPI wine_vk_init(INIT_ONCE *once, void *param, void **context) { HDC hdc; hdc = GetDC(0); vk_funcs = __wine_get_vulkan_driver(hdc, WINE_VULKAN_DRIVER_VERSION); ReleaseDC(0, hdc); if (!vk_funcs) ERR("Failed to load Wine graphics driver supporting Vulkan.\n"); else p_vkEnumerateInstanceVersion = vk_funcs->p_vkGetInstanceProcAddr(NULL, "vkEnumerateInstanceVersion"); return TRUE; } static void wine_vk_init_once(void) { static INIT_ONCE init_once = INIT_ONCE_STATIC_INIT; InitOnceExecuteOnce(&init_once, wine_vk_init, NULL, NULL); } /* Helper function for converting between win32 and host compatible VkInstanceCreateInfo. * This function takes care of extensions handled at winevulkan layer, a Wine graphics * driver is responsible for handling e.g. surface extensions. */ static VkResult wine_vk_instance_convert_create_info(const VkInstanceCreateInfo *src, VkInstanceCreateInfo *dst) { unsigned int i; VkResult res; *dst = *src; if ((res = convert_VkInstanceCreateInfo_struct_chain(src->pNext, dst)) < 0) { WARN("Failed to convert VkInstanceCreateInfo pNext chain, res=%d.\n", res); return res; } /* ICDs don't support any layers, so nothing to copy. Modern versions of the loader * filter this data out as well. */ dst->enabledLayerCount = 0; dst->ppEnabledLayerNames = NULL; TRACE("Enabled %u instance extensions.\n", dst->enabledExtensionCount); for (i = 0; i < dst->enabledExtensionCount; i++) { const char *extension_name = dst->ppEnabledExtensionNames[i]; TRACE("Extension %u: %s.\n", i, debugstr_a(extension_name)); if (!wine_vk_instance_extension_supported(extension_name)) { WARN("Extension %s is not supported.\n", debugstr_a(extension_name)); free_VkInstanceCreateInfo_struct_chain(dst); return VK_ERROR_EXTENSION_NOT_PRESENT; } } return VK_SUCCESS; } /* Helper function which stores wrapped physical devices in the instance object. */ static VkResult wine_vk_instance_load_physical_devices(struct VkInstance_T *instance) { VkPhysicalDevice *tmp_phys_devs; uint32_t phys_dev_count; unsigned int i; VkResult res; res = instance->funcs.p_vkEnumeratePhysicalDevices(instance->instance, &phys_dev_count, NULL); if (res != VK_SUCCESS) { ERR("Failed to enumerate physical devices, res=%d\n", res); return res; } if (!phys_dev_count) return res; if (!(tmp_phys_devs = heap_calloc(phys_dev_count, sizeof(*tmp_phys_devs)))) return VK_ERROR_OUT_OF_HOST_MEMORY; res = instance->funcs.p_vkEnumeratePhysicalDevices(instance->instance, &phys_dev_count, tmp_phys_devs); if (res != VK_SUCCESS) { heap_free(tmp_phys_devs); return res; } instance->phys_devs = heap_calloc(phys_dev_count, sizeof(*instance->phys_devs)); if (!instance->phys_devs) { heap_free(tmp_phys_devs); return VK_ERROR_OUT_OF_HOST_MEMORY; } /* Wrap each native physical device handle into a dispatchable object for the ICD loader. */ for (i = 0; i < phys_dev_count; i++) { struct VkPhysicalDevice_T *phys_dev = wine_vk_physical_device_alloc(instance, tmp_phys_devs[i]); if (!phys_dev) { ERR("Unable to allocate memory for physical device!\n"); heap_free(tmp_phys_devs); return VK_ERROR_OUT_OF_HOST_MEMORY; } instance->phys_devs[i] = phys_dev; instance->phys_dev_count = i + 1; } instance->phys_dev_count = phys_dev_count; heap_free(tmp_phys_devs); return VK_SUCCESS; } static struct VkPhysicalDevice_T *wine_vk_instance_wrap_physical_device(struct VkInstance_T *instance, VkPhysicalDevice physical_device) { unsigned int i; for (i = 0; i < instance->phys_dev_count; ++i) { struct VkPhysicalDevice_T *current = instance->phys_devs[i]; if (current->phys_dev == physical_device) return current; } ERR("Unrecognized physical device %p.\n", physical_device); return NULL; } /* Helper function used for freeing an instance structure. This function supports full * and partial object cleanups and can thus be used for vkCreateInstance failures. */ static void wine_vk_instance_free(struct VkInstance_T *instance) { if (!instance) return; if (instance->phys_devs) { unsigned int i; for (i = 0; i < instance->phys_dev_count; i++) { wine_vk_physical_device_free(instance->phys_devs[i]); } heap_free(instance->phys_devs); } if (instance->instance) vk_funcs->p_vkDestroyInstance(instance->instance, NULL /* allocator */); heap_free(instance); } VkResult WINAPI wine_vkAllocateCommandBuffers(VkDevice device, const VkCommandBufferAllocateInfo *allocate_info, VkCommandBuffer *buffers) { struct wine_cmd_pool *pool; VkResult res = VK_SUCCESS; unsigned int i; TRACE("%p, %p, %p\n", device, allocate_info, buffers); pool = wine_cmd_pool_from_handle(allocate_info->commandPool); memset(buffers, 0, allocate_info->commandBufferCount * sizeof(*buffers)); for (i = 0; i < allocate_info->commandBufferCount; i++) { #if defined(USE_STRUCT_CONVERSION) VkCommandBufferAllocateInfo_host allocate_info_host; #else VkCommandBufferAllocateInfo allocate_info_host; #endif /* TODO: future extensions (none yet) may require pNext conversion. */ allocate_info_host.pNext = allocate_info->pNext; allocate_info_host.sType = allocate_info->sType; allocate_info_host.commandPool = pool->command_pool; allocate_info_host.level = allocate_info->level; allocate_info_host.commandBufferCount = 1; TRACE("Allocating command buffer %u from pool 0x%s.\n", i, wine_dbgstr_longlong(allocate_info_host.commandPool)); if (!(buffers[i] = heap_alloc_zero(sizeof(**buffers)))) { res = VK_ERROR_OUT_OF_HOST_MEMORY; break; } buffers[i]->base.loader_magic = VULKAN_ICD_MAGIC_VALUE; buffers[i]->device = device; list_add_tail(&pool->command_buffers, &buffers[i]->pool_link); res = device->funcs.p_vkAllocateCommandBuffers(device->device, &allocate_info_host, &buffers[i]->command_buffer); if (res != VK_SUCCESS) { ERR("Failed to allocate command buffer, res=%d.\n", res); buffers[i]->command_buffer = VK_NULL_HANDLE; break; } } if (res != VK_SUCCESS) { wine_vk_free_command_buffers(device, pool, i + 1, buffers); memset(buffers, 0, allocate_info->commandBufferCount * sizeof(*buffers)); } return res; } void WINAPI wine_vkCmdExecuteCommands(VkCommandBuffer buffer, uint32_t count, const VkCommandBuffer *buffers) { VkCommandBuffer *tmp_buffers; unsigned int i; TRACE("%p %u %p\n", buffer, count, buffers); if (!buffers || !count) return; /* Unfortunately we need a temporary buffer as our command buffers are wrapped. * This call is called often and if a performance concern, we may want to use * alloca as we shouldn't need much memory and it needs to be cleaned up after * the call anyway. */ if (!(tmp_buffers = heap_alloc(count * sizeof(*tmp_buffers)))) { ERR("Failed to allocate memory for temporary command buffers\n"); return; } for (i = 0; i < count; i++) tmp_buffers[i] = buffers[i]->command_buffer; buffer->device->funcs.p_vkCmdExecuteCommands(buffer->command_buffer, count, tmp_buffers); heap_free(tmp_buffers); } VkResult WINAPI wine_vkCreateDevice(VkPhysicalDevice phys_dev, const VkDeviceCreateInfo *create_info, const VkAllocationCallbacks *allocator, VkDevice *device) { VkDeviceCreateInfo create_info_host; uint32_t max_queue_families; struct VkDevice_T *object; unsigned int i; VkResult res; TRACE("%p, %p, %p, %p\n", phys_dev, create_info, allocator, device); if (allocator) FIXME("Support for allocation callbacks not implemented yet\n"); if (TRACE_ON(vulkan)) { VkPhysicalDeviceProperties properties; wine_vkGetPhysicalDeviceProperties(phys_dev, &properties); TRACE("Device name: %s.\n", debugstr_a(properties.deviceName)); TRACE("Vendor ID: %#x, Device ID: %#x.\n", properties.vendorID, properties.deviceID); TRACE("Driver version: %#x.\n", properties.driverVersion); } if (!(object = heap_alloc_zero(sizeof(*object)))) return VK_ERROR_OUT_OF_HOST_MEMORY; object->base.loader_magic = VULKAN_ICD_MAGIC_VALUE; res = wine_vk_device_convert_create_info(create_info, &create_info_host); if (res != VK_SUCCESS) goto fail; res = phys_dev->instance->funcs.p_vkCreateDevice(phys_dev->phys_dev, &create_info_host, NULL /* allocator */, &object->device); wine_vk_device_free_create_info(&create_info_host); if (res != VK_SUCCESS) { WARN("Failed to create device, res=%d.\n", res); goto fail; } /* Just load all function pointers we are aware off. The loader takes care of filtering. * We use vkGetDeviceProcAddr as opposed to vkGetInstanceProcAddr for efficiency reasons * as functions pass through fewer dispatch tables within the loader. */ #define USE_VK_FUNC(name) \ object->funcs.p_##name = (void *)vk_funcs->p_vkGetDeviceProcAddr(object->device, #name); \ if (object->funcs.p_##name == NULL) \ TRACE("Not found '%s'.\n", #name); ALL_VK_DEVICE_FUNCS() #undef USE_VK_FUNC /* We need to cache all queues within the device as each requires wrapping since queues are * dispatchable objects. */ phys_dev->instance->funcs.p_vkGetPhysicalDeviceQueueFamilyProperties(phys_dev->phys_dev, &max_queue_families, NULL); object->max_queue_families = max_queue_families; TRACE("Max queue families: %u.\n", object->max_queue_families); if (!(object->queues = heap_calloc(max_queue_families, sizeof(*object->queues)))) { res = VK_ERROR_OUT_OF_HOST_MEMORY; goto fail; } for (i = 0; i < create_info_host.queueCreateInfoCount; i++) { uint32_t flags = create_info_host.pQueueCreateInfos[i].flags; uint32_t family_index = create_info_host.pQueueCreateInfos[i].queueFamilyIndex; uint32_t queue_count = create_info_host.pQueueCreateInfos[i].queueCount; TRACE("Queue family index %u, queue count %u.\n", family_index, queue_count); if (!(object->queues[family_index] = wine_vk_device_alloc_queues(object, family_index, queue_count, flags))) { ERR("Failed to allocate memory for queues.\n"); res = VK_ERROR_OUT_OF_HOST_MEMORY; goto fail; } } object->quirks = phys_dev->instance->quirks; *device = object; TRACE("Created device %p (native device %p).\n", object, object->device); return VK_SUCCESS; fail: wine_vk_device_free(object); return res; } VkResult WINAPI wine_vkCreateInstance(const VkInstanceCreateInfo *create_info, const VkAllocationCallbacks *allocator, VkInstance *instance) { VkInstanceCreateInfo create_info_host; const VkApplicationInfo *app_info; struct VkInstance_T *object; VkResult res; TRACE("create_info %p, allocator %p, instance %p\n", create_info, allocator, instance); wine_vk_init_once(); if (!vk_funcs) return VK_ERROR_INITIALIZATION_FAILED; if (allocator) FIXME("Support for allocation callbacks not implemented yet\n"); if (!(object = heap_alloc_zero(sizeof(*object)))) { ERR("Failed to allocate memory for instance\n"); return VK_ERROR_OUT_OF_HOST_MEMORY; } object->base.loader_magic = VULKAN_ICD_MAGIC_VALUE; res = wine_vk_instance_convert_create_info(create_info, &create_info_host); if (res != VK_SUCCESS) { wine_vk_instance_free(object); return res; } res = vk_funcs->p_vkCreateInstance(&create_info_host, NULL /* allocator */, &object->instance); free_VkInstanceCreateInfo_struct_chain(&create_info_host); if (res != VK_SUCCESS) { ERR("Failed to create instance, res=%d\n", res); wine_vk_instance_free(object); return res; } /* Load all instance functions we are aware of. Note the loader takes care * of any filtering for extensions which were not requested, but which the * ICD may support. */ #define USE_VK_FUNC(name) \ object->funcs.p_##name = (void *)vk_funcs->p_vkGetInstanceProcAddr(object->instance, #name); ALL_VK_INSTANCE_FUNCS() #undef USE_VK_FUNC /* Cache physical devices for vkEnumeratePhysicalDevices within the instance as * each vkPhysicalDevice is a dispatchable object, which means we need to wrap * the native physical devices and present those to the application. * Cleanup happens as part of wine_vkDestroyInstance. */ res = wine_vk_instance_load_physical_devices(object); if (res != VK_SUCCESS) { ERR("Failed to load physical devices, res=%d\n", res); wine_vk_instance_free(object); return res; } if ((app_info = create_info->pApplicationInfo)) { TRACE("Application name %s, application version %#x.\n", debugstr_a(app_info->pApplicationName), app_info->applicationVersion); TRACE("Engine name %s, engine version %#x.\n", debugstr_a(app_info->pEngineName), app_info->engineVersion); TRACE("API version %#x.\n", app_info->apiVersion); if (app_info->pEngineName && !strcmp(app_info->pEngineName, "idTech")) object->quirks |= WINEVULKAN_QUIRK_GET_DEVICE_PROC_ADDR; } object->quirks |= WINEVULKAN_QUIRK_ADJUST_MAX_IMAGE_COUNT; *instance = object; TRACE("Created instance %p (native instance %p).\n", object, object->instance); return VK_SUCCESS; } void WINAPI wine_vkDestroyDevice(VkDevice device, const VkAllocationCallbacks *allocator) { TRACE("%p %p\n", device, allocator); if (allocator) FIXME("Support for allocation callbacks not implemented yet\n"); wine_vk_device_free(device); } void WINAPI wine_vkDestroyInstance(VkInstance instance, const VkAllocationCallbacks *allocator) { TRACE("%p, %p\n", instance, allocator); if (allocator) FIXME("Support allocation allocators\n"); wine_vk_instance_free(instance); } VkResult WINAPI wine_vkEnumerateDeviceExtensionProperties(VkPhysicalDevice phys_dev, const char *layer_name, uint32_t *count, VkExtensionProperties *properties) { TRACE("%p, %p, %p, %p\n", phys_dev, layer_name, count, properties); /* This shouldn't get called with layer_name set, the ICD loader prevents it. */ if (layer_name) { ERR("Layer enumeration not supported from ICD.\n"); return VK_ERROR_LAYER_NOT_PRESENT; } if (!properties) { *count = phys_dev->extension_count; return VK_SUCCESS; } *count = min(*count, phys_dev->extension_count); memcpy(properties, phys_dev->extensions, *count * sizeof(*properties)); TRACE("Returning %u extensions.\n", *count); return *count < phys_dev->extension_count ? VK_INCOMPLETE : VK_SUCCESS; } VkResult WINAPI wine_vkEnumerateInstanceExtensionProperties(const char *layer_name, uint32_t *count, VkExtensionProperties *properties) { uint32_t num_properties = 0, num_host_properties; VkExtensionProperties *host_properties; unsigned int i, j; VkResult res; TRACE("%p, %p, %p\n", layer_name, count, properties); if (layer_name) { WARN("Layer enumeration not supported from ICD.\n"); return VK_ERROR_LAYER_NOT_PRESENT; } wine_vk_init_once(); if (!vk_funcs) return VK_ERROR_INITIALIZATION_FAILED; res = vk_funcs->p_vkEnumerateInstanceExtensionProperties(NULL, &num_host_properties, NULL); if (res != VK_SUCCESS) return res; if (!(host_properties = heap_calloc(num_host_properties, sizeof(*host_properties)))) return VK_ERROR_OUT_OF_HOST_MEMORY; res = vk_funcs->p_vkEnumerateInstanceExtensionProperties(NULL, &num_host_properties, host_properties); if (res != VK_SUCCESS) { ERR("Failed to retrieve host properties, res=%d.\n", res); heap_free(host_properties); return res; } /* The Wine graphics driver provides us with all extensions supported by the host side * including extension fixup (e.g. VK_KHR_xlib_surface -> VK_KHR_win32_surface). It is * up to us here to filter the list down to extensions for which we have thunks. */ for (i = 0; i < num_host_properties; i++) { if (wine_vk_instance_extension_supported(host_properties[i].extensionName)) num_properties++; else TRACE("Instance extension '%s' is not supported.\n", host_properties[i].extensionName); } if (!properties) { TRACE("Returning %u extensions.\n", num_properties); *count = num_properties; heap_free(host_properties); return VK_SUCCESS; } for (i = 0, j = 0; i < num_host_properties && j < *count; i++) { if (wine_vk_instance_extension_supported(host_properties[i].extensionName)) { TRACE("Enabling extension '%s'.\n", host_properties[i].extensionName); properties[j++] = host_properties[i]; } } *count = min(*count, num_properties); heap_free(host_properties); return *count < num_properties ? VK_INCOMPLETE : VK_SUCCESS; } VkResult WINAPI wine_vkEnumerateInstanceLayerProperties(uint32_t *count, VkLayerProperties *properties) { TRACE("%p, %p\n", count, properties); if (!properties) { *count = 0; return VK_SUCCESS; } return VK_ERROR_LAYER_NOT_PRESENT; } VkResult WINAPI wine_vkEnumerateInstanceVersion(uint32_t *version) { VkResult res; TRACE("%p\n", version); wine_vk_init_once(); if (p_vkEnumerateInstanceVersion) { res = p_vkEnumerateInstanceVersion(version); } else { *version = VK_API_VERSION_1_0; res = VK_SUCCESS; } TRACE("API version %u.%u.%u.\n", VK_VERSION_MAJOR(*version), VK_VERSION_MINOR(*version), VK_VERSION_PATCH(*version)); *version = min(WINE_VK_VERSION, *version); return res; } VkResult WINAPI wine_vkEnumeratePhysicalDevices(VkInstance instance, uint32_t *count, VkPhysicalDevice *devices) { unsigned int i; TRACE("%p %p %p\n", instance, count, devices); if (!devices) { *count = instance->phys_dev_count; return VK_SUCCESS; } *count = min(*count, instance->phys_dev_count); for (i = 0; i < *count; i++) { devices[i] = instance->phys_devs[i]; } TRACE("Returning %u devices.\n", *count); return *count < instance->phys_dev_count ? VK_INCOMPLETE : VK_SUCCESS; } void WINAPI wine_vkFreeCommandBuffers(VkDevice device, VkCommandPool pool_handle, uint32_t count, const VkCommandBuffer *buffers) { struct wine_cmd_pool *pool = wine_cmd_pool_from_handle(pool_handle); TRACE("%p, 0x%s, %u, %p\n", device, wine_dbgstr_longlong(pool_handle), count, buffers); wine_vk_free_command_buffers(device, pool, count, buffers); } PFN_vkVoidFunction WINAPI wine_vkGetDeviceProcAddr(VkDevice device, const char *name) { void *func; TRACE("%p, %s\n", device, debugstr_a(name)); /* The spec leaves return value undefined for a NULL device, let's just return NULL. */ if (!device || !name) return NULL; /* Per the spec, we are only supposed to return device functions as in functions * for which the first parameter is vkDevice or a child of vkDevice like a * vkCommandBuffer or vkQueue. * Loader takes care of filtering of extensions which are enabled or not. */ func = wine_vk_get_device_proc_addr(name); if (func) return func; /* vkGetDeviceProcAddr was intended for loading device and subdevice functions. * idTech 6 titles such as Doom and Wolfenstein II, however use it also for * loading of instance functions. This is undefined behavior as the specification * disallows using any of the returned function pointers outside of device / * subdevice objects. The games don't actually use the function pointers and if they * did, they would crash as VkInstance / VkPhysicalDevice parameters need unwrapping. * Khronos clarified behavior in the Vulkan spec and expects drivers to get updated, * however it would require both driver and game fixes. * https://github.com/KhronosGroup/Vulkan-LoaderAndValidationLayers/issues/2323 * https://github.com/KhronosGroup/Vulkan-Docs/issues/655 */ if (device->quirks & WINEVULKAN_QUIRK_GET_DEVICE_PROC_ADDR && (func = wine_vk_get_instance_proc_addr(name))) { WARN("Returning instance function %s.\n", debugstr_a(name)); return func; } WARN("Unsupported device function: %s.\n", debugstr_a(name)); return NULL; } void WINAPI wine_vkGetDeviceQueue(VkDevice device, uint32_t family_index, uint32_t queue_index, VkQueue *queue) { TRACE("%p, %u, %u, %p\n", device, family_index, queue_index, queue); *queue = &device->queues[family_index][queue_index]; } void WINAPI wine_vkGetDeviceQueue2(VkDevice device, const VkDeviceQueueInfo2 *info, VkQueue *queue) { struct VkQueue_T *matching_queue; const VkBaseInStructure *chain; TRACE("%p, %p, %p\n", device, info, queue); if ((chain = info->pNext)) FIXME("Ignoring a linked structure of type %u.\n", chain->sType); matching_queue = &device->queues[info->queueFamilyIndex][info->queueIndex]; if (matching_queue->flags != info->flags) { WARN("No matching flags were specified %#x, %#x.\n", matching_queue->flags, info->flags); matching_queue = VK_NULL_HANDLE; } *queue = matching_queue; } PFN_vkVoidFunction WINAPI wine_vkGetInstanceProcAddr(VkInstance instance, const char *name) { void *func; TRACE("%p, %s\n", instance, debugstr_a(name)); if (!name) return NULL; /* vkGetInstanceProcAddr can load most Vulkan functions when an instance is passed in, however * for a NULL instance it can only load global functions. */ func = wine_vk_get_global_proc_addr(name); if (func) { return func; } if (!instance) { WARN("Global function %s not found.\n", debugstr_a(name)); return NULL; } func = wine_vk_get_instance_proc_addr(name); if (func) return func; /* vkGetInstanceProcAddr also loads any children of instance, so device functions as well. */ func = wine_vk_get_device_proc_addr(name); if (func) return func; WARN("Unsupported device or instance function: %s.\n", debugstr_a(name)); return NULL; } void * WINAPI wine_vk_icdGetInstanceProcAddr(VkInstance instance, const char *name) { TRACE("%p, %s\n", instance, debugstr_a(name)); /* Initial version of the Vulkan ICD spec required vkGetInstanceProcAddr to be * exported. vk_icdGetInstanceProcAddr was added later to separate ICD calls from * Vulkan API. One of them in our case should forward to the other, so just forward * to the older vkGetInstanceProcAddr. */ return wine_vkGetInstanceProcAddr(instance, name); } VkResult WINAPI wine_vk_icdNegotiateLoaderICDInterfaceVersion(uint32_t *supported_version) { uint32_t req_version; TRACE("%p\n", supported_version); /* The spec is not clear how to handle this. Mesa drivers don't check, but it * is probably best to not explode. VK_INCOMPLETE seems to be the closest value. */ if (!supported_version) return VK_INCOMPLETE; req_version = *supported_version; *supported_version = min(req_version, WINE_VULKAN_ICD_VERSION); TRACE("Loader requested ICD version %u, returning %u\n", req_version, *supported_version); return VK_SUCCESS; } VkResult WINAPI wine_vkQueueSubmit(VkQueue queue, uint32_t count, const VkSubmitInfo *submits, VkFence fence) { VkSubmitInfo *submits_host; VkResult res; VkCommandBuffer *command_buffers; unsigned int i, j, num_command_buffers; TRACE("%p %u %p 0x%s\n", queue, count, submits, wine_dbgstr_longlong(fence)); if (count == 0) { return queue->device->funcs.p_vkQueueSubmit(queue->queue, 0, NULL, fence); } submits_host = heap_calloc(count, sizeof(*submits_host)); if (!submits_host) { ERR("Unable to allocate memory for submit buffers!\n"); return VK_ERROR_OUT_OF_HOST_MEMORY; } for (i = 0; i < count; i++) { memcpy(&submits_host[i], &submits[i], sizeof(*submits_host)); num_command_buffers = submits[i].commandBufferCount; command_buffers = heap_calloc(num_command_buffers, sizeof(*submits_host)); if (!command_buffers) { ERR("Unable to allocate memory for command buffers!\n"); res = VK_ERROR_OUT_OF_HOST_MEMORY; goto done; } for (j = 0; j < num_command_buffers; j++) { command_buffers[j] = submits[i].pCommandBuffers[j]->command_buffer; } submits_host[i].pCommandBuffers = command_buffers; } res = queue->device->funcs.p_vkQueueSubmit(queue->queue, count, submits_host, fence); done: for (i = 0; i < count; i++) { heap_free((void *)submits_host[i].pCommandBuffers); } heap_free(submits_host); TRACE("Returning %d\n", res); return res; } VkResult WINAPI wine_vkCreateCommandPool(VkDevice device, const VkCommandPoolCreateInfo *info, const VkAllocationCallbacks *allocator, VkCommandPool *command_pool) { struct wine_cmd_pool *object; VkResult res; TRACE("%p, %p, %p, %p\n", device, info, allocator, command_pool); if (allocator) FIXME("Support for allocation callbacks not implemented yet\n"); if (!(object = heap_alloc_zero(sizeof(*object)))) return VK_ERROR_OUT_OF_HOST_MEMORY; list_init(&object->command_buffers); res = device->funcs.p_vkCreateCommandPool(device->device, info, NULL, &object->command_pool); if (res == VK_SUCCESS) *command_pool = wine_cmd_pool_to_handle(object); else heap_free(object); return res; } void WINAPI wine_vkDestroyCommandPool(VkDevice device, VkCommandPool handle, const VkAllocationCallbacks *allocator) { struct wine_cmd_pool *pool = wine_cmd_pool_from_handle(handle); struct VkCommandBuffer_T *buffer, *cursor; TRACE("%p, 0x%s, %p\n", device, wine_dbgstr_longlong(handle), allocator); if (!handle) return; if (allocator) FIXME("Support for allocation callbacks not implemented yet\n"); /* The Vulkan spec says: * * "When a pool is destroyed, all command buffers allocated from the pool are freed." */ LIST_FOR_EACH_ENTRY_SAFE(buffer, cursor, &pool->command_buffers, struct VkCommandBuffer_T, pool_link) { heap_free(buffer); } device->funcs.p_vkDestroyCommandPool(device->device, pool->command_pool, NULL); heap_free(pool); } static VkResult wine_vk_enumerate_physical_device_groups(struct VkInstance_T *instance, VkResult (*p_vkEnumeratePhysicalDeviceGroups)(VkInstance, uint32_t *, VkPhysicalDeviceGroupProperties *), uint32_t *count, VkPhysicalDeviceGroupProperties *properties) { unsigned int i, j; VkResult res; res = p_vkEnumeratePhysicalDeviceGroups(instance->instance, count, properties); if (res < 0 || !properties) return res; for (i = 0; i < *count; ++i) { VkPhysicalDeviceGroupProperties *current = &properties[i]; for (j = 0; j < current->physicalDeviceCount; ++j) { VkPhysicalDevice dev = current->physicalDevices[j]; if (!(current->physicalDevices[j] = wine_vk_instance_wrap_physical_device(instance, dev))) return VK_ERROR_INITIALIZATION_FAILED; } } return res; } VkResult WINAPI wine_vkEnumeratePhysicalDeviceGroups(VkInstance instance, uint32_t *count, VkPhysicalDeviceGroupProperties *properties) { TRACE("%p, %p, %p\n", instance, count, properties); return wine_vk_enumerate_physical_device_groups(instance, instance->funcs.p_vkEnumeratePhysicalDeviceGroups, count, properties); } VkResult WINAPI wine_vkEnumeratePhysicalDeviceGroupsKHR(VkInstance instance, uint32_t *count, VkPhysicalDeviceGroupProperties *properties) { TRACE("%p, %p, %p\n", instance, count, properties); return wine_vk_enumerate_physical_device_groups(instance, instance->funcs.p_vkEnumeratePhysicalDeviceGroupsKHR, count, properties); } void WINAPI wine_vkGetPhysicalDeviceExternalFenceProperties(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalFenceInfo *fence_info, VkExternalFenceProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, fence_info, properties); properties->exportFromImportedHandleTypes = 0; properties->compatibleHandleTypes = 0; properties->externalFenceFeatures = 0; } void WINAPI wine_vkGetPhysicalDeviceExternalFencePropertiesKHR(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalFenceInfo *fence_info, VkExternalFenceProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, fence_info, properties); properties->exportFromImportedHandleTypes = 0; properties->compatibleHandleTypes = 0; properties->externalFenceFeatures = 0; } void WINAPI wine_vkGetPhysicalDeviceExternalBufferProperties(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalBufferInfo *buffer_info, VkExternalBufferProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, buffer_info, properties); memset(&properties->externalMemoryProperties, 0, sizeof(properties->externalMemoryProperties)); } void WINAPI wine_vkGetPhysicalDeviceExternalBufferPropertiesKHR(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalBufferInfo *buffer_info, VkExternalBufferProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, buffer_info, properties); memset(&properties->externalMemoryProperties, 0, sizeof(properties->externalMemoryProperties)); } VkResult WINAPI wine_vkGetPhysicalDeviceImageFormatProperties2(VkPhysicalDevice phys_dev, const VkPhysicalDeviceImageFormatInfo2 *format_info, VkImageFormatProperties2 *properties) { VkExternalImageFormatProperties *external_image_properties; VkResult res; TRACE("%p, %p, %p\n", phys_dev, format_info, properties); res = thunk_vkGetPhysicalDeviceImageFormatProperties2(phys_dev, format_info, properties); if ((external_image_properties = wine_vk_find_struct(properties, EXTERNAL_IMAGE_FORMAT_PROPERTIES))) { VkExternalMemoryProperties *p = &external_image_properties->externalMemoryProperties; p->externalMemoryFeatures = 0; p->exportFromImportedHandleTypes = 0; p->compatibleHandleTypes = 0; } return res; } VkResult WINAPI wine_vkGetPhysicalDeviceImageFormatProperties2KHR(VkPhysicalDevice phys_dev, const VkPhysicalDeviceImageFormatInfo2 *format_info, VkImageFormatProperties2 *properties) { VkExternalImageFormatProperties *external_image_properties; VkResult res; TRACE("%p, %p, %p\n", phys_dev, format_info, properties); res = thunk_vkGetPhysicalDeviceImageFormatProperties2KHR(phys_dev, format_info, properties); if ((external_image_properties = wine_vk_find_struct(properties, EXTERNAL_IMAGE_FORMAT_PROPERTIES))) { VkExternalMemoryProperties *p = &external_image_properties->externalMemoryProperties; p->externalMemoryFeatures = 0; p->exportFromImportedHandleTypes = 0; p->compatibleHandleTypes = 0; } return res; } /* From ntdll/unix/sync.c */ #define NANOSECONDS_IN_A_SECOND 1000000000 #define TICKSPERSEC 10000000 static inline VkTimeDomainEXT get_performance_counter_time_domain(void) { #if !defined(__APPLE__) && defined(HAVE_CLOCK_GETTIME) # ifdef CLOCK_MONOTONIC_RAW return VK_TIME_DOMAIN_CLOCK_MONOTONIC_RAW_EXT; # else return VK_TIME_DOMAIN_CLOCK_MONOTONIC_EXT; # endif #else FIXME("No mapping for VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT on this platform."); return VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT; #endif } static VkTimeDomainEXT map_to_host_time_domain(VkTimeDomainEXT domain) { /* Matches ntdll/unix/sync.c's performance counter implementation. */ if (domain == VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT) return get_performance_counter_time_domain(); return domain; } static inline uint64_t convert_monotonic_timestamp(uint64_t value) { return value / (NANOSECONDS_IN_A_SECOND / TICKSPERSEC); } static inline uint64_t convert_timestamp(VkTimeDomainEXT host_domain, VkTimeDomainEXT target_domain, uint64_t value) { if (host_domain == target_domain) return value; /* Convert between MONOTONIC time in ns -> QueryPerformanceCounter */ if ((host_domain == VK_TIME_DOMAIN_CLOCK_MONOTONIC_RAW_EXT || host_domain == VK_TIME_DOMAIN_CLOCK_MONOTONIC_EXT) && target_domain == VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT) return convert_monotonic_timestamp(value); FIXME("Couldn't translate between host domain %d and target domain %d", host_domain, target_domain); return value; } VkResult WINAPI wine_vkGetCalibratedTimestampsEXT(VkDevice device, uint32_t timestamp_count, const VkCalibratedTimestampInfoEXT *timestamp_infos, uint64_t *timestamps, uint64_t *max_deviation) { VkCalibratedTimestampInfoEXT* host_timestamp_infos; unsigned int i; VkResult res; TRACE("%p, %u, %p, %p, %p\n", device, timestamp_count, timestamp_infos, timestamps, max_deviation); if (!(host_timestamp_infos = heap_alloc(sizeof(VkCalibratedTimestampInfoEXT) * timestamp_count))) return VK_ERROR_OUT_OF_HOST_MEMORY; for (i = 0; i < timestamp_count; i++) { host_timestamp_infos[i].sType = timestamp_infos[i].sType; host_timestamp_infos[i].pNext = timestamp_infos[i].pNext; host_timestamp_infos[i].timeDomain = map_to_host_time_domain(timestamp_infos[i].timeDomain); } res = device->funcs.p_vkGetCalibratedTimestampsEXT(device->device, timestamp_count, host_timestamp_infos, timestamps, max_deviation); if (res != VK_SUCCESS) return res; for (i = 0; i < timestamp_count; i++) timestamps[i] = convert_timestamp(host_timestamp_infos[i].timeDomain, timestamp_infos[i].timeDomain, timestamps[i]); heap_free(host_timestamp_infos); return res; } VkResult WINAPI wine_vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(VkPhysicalDevice phys_dev, uint32_t *time_domain_count, VkTimeDomainEXT *time_domains) { BOOL supports_device = FALSE, supports_monotonic = FALSE, supports_monotonic_raw = FALSE; const VkTimeDomainEXT performance_counter_domain = get_performance_counter_time_domain(); VkTimeDomainEXT *host_time_domains; uint32_t host_time_domain_count; VkTimeDomainEXT out_time_domains[2]; uint32_t out_time_domain_count; unsigned int i; VkResult res; TRACE("%p, %p, %p\n", phys_dev, time_domain_count, time_domains); /* Find out the time domains supported on the host */ res = phys_dev->instance->funcs.p_vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(phys_dev->phys_dev, &host_time_domain_count, NULL); if (res != VK_SUCCESS) return res; if (!(host_time_domains = heap_alloc(sizeof(VkTimeDomainEXT) * host_time_domain_count))) return VK_ERROR_OUT_OF_HOST_MEMORY; res = phys_dev->instance->funcs.p_vkGetPhysicalDeviceCalibrateableTimeDomainsEXT(phys_dev->phys_dev, &host_time_domain_count, host_time_domains); if (res != VK_SUCCESS) { heap_free(host_time_domains); return res; } for (i = 0; i < host_time_domain_count; i++) { if (host_time_domains[i] == VK_TIME_DOMAIN_DEVICE_EXT) supports_device = TRUE; else if (host_time_domains[i] == VK_TIME_DOMAIN_CLOCK_MONOTONIC_EXT) supports_monotonic = TRUE; else if (host_time_domains[i] == VK_TIME_DOMAIN_CLOCK_MONOTONIC_RAW_EXT) supports_monotonic_raw = TRUE; else FIXME("Unknown time domain %d", host_time_domains[i]); } heap_free(host_time_domains); out_time_domain_count = 0; /* Map our monotonic times -> QPC */ if (supports_monotonic_raw && performance_counter_domain == VK_TIME_DOMAIN_CLOCK_MONOTONIC_RAW_EXT) out_time_domains[out_time_domain_count++] = VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT; else if (supports_monotonic && performance_counter_domain == VK_TIME_DOMAIN_CLOCK_MONOTONIC_EXT) out_time_domains[out_time_domain_count++] = VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT; else FIXME("VK_TIME_DOMAIN_QUERY_PERFORMANCE_COUNTER_EXT not supported on this platform."); /* Forward the device domain time */ if (supports_device) out_time_domains[out_time_domain_count++] = VK_TIME_DOMAIN_DEVICE_EXT; /* Send the count/domains back to the app */ if (!time_domains) { *time_domain_count = out_time_domain_count; return VK_SUCCESS; } for (i = 0; i < min(*time_domain_count, out_time_domain_count); i++) time_domains[i] = out_time_domains[i]; res = *time_domain_count < out_time_domain_count ? VK_INCOMPLETE : VK_SUCCESS; *time_domain_count = out_time_domain_count; return res; } static HANDLE get_display_device_init_mutex(void) { static const WCHAR init_mutexW[] = {'d','i','s','p','l','a','y','_','d','e','v','i','c','e','_','i','n','i','t',0}; HANDLE mutex = CreateMutexW(NULL, FALSE, init_mutexW); WaitForSingleObject(mutex, INFINITE); return mutex; } static void release_display_device_init_mutex(HANDLE mutex) { ReleaseMutex(mutex); CloseHandle(mutex); } /* Wait until graphics driver is loaded by explorer */ static void wait_graphics_driver_ready(void) { static BOOL ready = FALSE; if (!ready) { SendMessageW(GetDesktopWindow(), WM_NULL, 0, 0); ready = TRUE; } } static void fill_luid_property(VkPhysicalDeviceProperties2 *properties2) { static const WCHAR pci[] = {'P','C','I',0}; VkPhysicalDeviceIDProperties *id; SP_DEVINFO_DATA device_data; DWORD type, device_idx = 0; HDEVINFO devinfo; HANDLE mutex; GUID uuid; LUID luid; if (!(id = wine_vk_find_struct(properties2, PHYSICAL_DEVICE_ID_PROPERTIES))) return; wait_graphics_driver_ready(); mutex = get_display_device_init_mutex(); devinfo = SetupDiGetClassDevsW(&GUID_DEVCLASS_DISPLAY, pci, NULL, 0); device_data.cbSize = sizeof(device_data); while (SetupDiEnumDeviceInfo(devinfo, device_idx++, &device_data)) { if (!SetupDiGetDevicePropertyW(devinfo, &device_data, &WINE_DEVPROPKEY_GPU_VULKAN_UUID, &type, (BYTE *)&uuid, sizeof(uuid), NULL, 0)) continue; if (!IsEqualGUID(&uuid, id->deviceUUID)) continue; if (SetupDiGetDevicePropertyW(devinfo, &device_data, &DEVPROPKEY_GPU_LUID, &type, (BYTE *)&luid, sizeof(luid), NULL, 0)) { memcpy(&id->deviceLUID, &luid, sizeof(id->deviceLUID)); id->deviceLUIDValid = VK_TRUE; id->deviceNodeMask = 1; break; } } SetupDiDestroyDeviceInfoList(devinfo); release_display_device_init_mutex(mutex); TRACE("deviceName:%s deviceLUIDValid:%d LUID:%08x:%08x deviceNodeMask:%#x.\n", properties2->properties.deviceName, id->deviceLUIDValid, luid.HighPart, luid.LowPart, id->deviceNodeMask); } void WINAPI wine_vkGetPhysicalDeviceProperties2(VkPhysicalDevice phys_dev, VkPhysicalDeviceProperties2 *properties2) { TRACE("%p, %p\n", phys_dev, properties2); thunk_vkGetPhysicalDeviceProperties2(phys_dev, properties2); fill_luid_property(properties2); } void WINAPI wine_vkGetPhysicalDeviceProperties2KHR(VkPhysicalDevice phys_dev, VkPhysicalDeviceProperties2 *properties2) { TRACE("%p, %p\n", phys_dev, properties2); thunk_vkGetPhysicalDeviceProperties2KHR(phys_dev, properties2); fill_luid_property(properties2); } void WINAPI wine_vkGetPhysicalDeviceExternalSemaphoreProperties(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalSemaphoreInfo *semaphore_info, VkExternalSemaphoreProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, semaphore_info, properties); properties->exportFromImportedHandleTypes = 0; properties->compatibleHandleTypes = 0; properties->externalSemaphoreFeatures = 0; } void WINAPI wine_vkGetPhysicalDeviceExternalSemaphorePropertiesKHR(VkPhysicalDevice phys_dev, const VkPhysicalDeviceExternalSemaphoreInfo *semaphore_info, VkExternalSemaphoreProperties *properties) { TRACE("%p, %p, %p\n", phys_dev, semaphore_info, properties); properties->exportFromImportedHandleTypes = 0; properties->compatibleHandleTypes = 0; properties->externalSemaphoreFeatures = 0; } static uint64_t unwrap_object_handle(VkObjectType type, uint64_t handle) { switch (type) { case VK_OBJECT_TYPE_DEVICE: return (uint64_t) (uintptr_t) ((VkDevice) (uintptr_t) handle)->device; case VK_OBJECT_TYPE_QUEUE: return (uint64_t) (uintptr_t) ((VkQueue) (uintptr_t) handle)->queue; case VK_OBJECT_TYPE_COMMAND_BUFFER: return (uint64_t) (uintptr_t) ((VkCommandBuffer) (uintptr_t) handle)->command_buffer; case VK_OBJECT_TYPE_COMMAND_POOL: return (uint64_t) wine_cmd_pool_from_handle(handle)->command_pool; default: return handle; } } VkResult WINAPI wine_vkSetPrivateDataEXT(VkDevice device, VkObjectType object_type, uint64_t object_handle, VkPrivateDataSlotEXT private_data_slot, uint64_t data) { TRACE("%p, %#x, 0x%s, 0x%s, 0x%s\n", device, object_type, wine_dbgstr_longlong(object_handle), wine_dbgstr_longlong(private_data_slot), wine_dbgstr_longlong(data)); object_handle = unwrap_object_handle(object_type, object_handle); return device->funcs.p_vkSetPrivateDataEXT(device->device, object_type, object_handle, private_data_slot, data); } void WINAPI wine_vkGetPrivateDataEXT(VkDevice device, VkObjectType object_type, uint64_t object_handle, VkPrivateDataSlotEXT private_data_slot, uint64_t *data) { TRACE("%p, %#x, 0x%s, 0x%s, %p\n", device, object_type, wine_dbgstr_longlong(object_handle), wine_dbgstr_longlong(private_data_slot), data); object_handle = unwrap_object_handle(object_type, object_handle); device->funcs.p_vkGetPrivateDataEXT(device->device, object_type, object_handle, private_data_slot, data); } static inline void adjust_max_image_count(VkPhysicalDevice phys_dev, VkSurfaceCapabilitiesKHR* capabilities) { /* Many Windows games, for example Strange Brigade, No Man's Sky, Path of Exile * and World War Z, do not expect that maxImageCount can be set to 0. * A value of 0 means that there is no limit on the number of images. * Nvidia reports 8 on Windows, AMD 16. * https://vulkan.gpuinfo.org/displayreport.php?id=9122#surface * https://vulkan.gpuinfo.org/displayreport.php?id=9121#surface */ if ((phys_dev->instance->quirks & WINEVULKAN_QUIRK_ADJUST_MAX_IMAGE_COUNT) && !capabilities->maxImageCount) { capabilities->maxImageCount = max(capabilities->minImageCount, 16); } } VkResult WINAPI wine_vkGetPhysicalDeviceSurfaceCapabilitiesKHR(VkPhysicalDevice phys_dev, VkSurfaceKHR surface, VkSurfaceCapabilitiesKHR *capabilities) { VkResult res; TRACE("%p, 0x%s, %p\n", phys_dev, wine_dbgstr_longlong(surface), capabilities); res = thunk_vkGetPhysicalDeviceSurfaceCapabilitiesKHR(phys_dev, surface, capabilities); if (res == VK_SUCCESS) adjust_max_image_count(phys_dev, capabilities); return res; } VkResult WINAPI wine_vkGetPhysicalDeviceSurfaceCapabilities2KHR(VkPhysicalDevice phys_dev, const VkPhysicalDeviceSurfaceInfo2KHR *surface_info, VkSurfaceCapabilities2KHR *capabilities) { VkResult res; TRACE("%p, %p, %p\n", phys_dev, surface_info, capabilities); res = thunk_vkGetPhysicalDeviceSurfaceCapabilities2KHR(phys_dev, surface_info, capabilities); if (res == VK_SUCCESS) adjust_max_image_count(phys_dev, &capabilities->surfaceCapabilities); return res; } BOOL WINAPI DllMain(HINSTANCE hinst, DWORD reason, void *reserved) { TRACE("%p, %u, %p\n", hinst, reason, reserved); switch (reason) { case DLL_PROCESS_ATTACH: hinstance = hinst; DisableThreadLibraryCalls(hinst); break; } return TRUE; } static const struct vulkan_func vk_global_dispatch_table[] = { /* These functions must call wine_vk_init_once() before accessing vk_funcs. */ {"vkCreateInstance", &wine_vkCreateInstance}, {"vkEnumerateInstanceExtensionProperties", &wine_vkEnumerateInstanceExtensionProperties}, {"vkEnumerateInstanceLayerProperties", &wine_vkEnumerateInstanceLayerProperties}, {"vkEnumerateInstanceVersion", &wine_vkEnumerateInstanceVersion}, {"vkGetInstanceProcAddr", &wine_vkGetInstanceProcAddr}, }; static void *wine_vk_get_global_proc_addr(const char *name) { unsigned int i; for (i = 0; i < ARRAY_SIZE(vk_global_dispatch_table); i++) { if (strcmp(name, vk_global_dispatch_table[i].name) == 0) { TRACE("Found name=%s in global table\n", debugstr_a(name)); return vk_global_dispatch_table[i].func; } } return NULL; } /* * Wrapper around driver vkGetInstanceProcAddr implementation. * Allows winelib applications to access Vulkan functions with Wine * additions and native ABI. */ void *native_vkGetInstanceProcAddrWINE(VkInstance instance, const char *name) { return vk_funcs->p_vkGetInstanceProcAddr(instance, name); } static const WCHAR winevulkan_json_resW[] = {'w','i','n','e','v','u','l','k','a','n','_','j','s','o','n',0}; static const WCHAR winevulkan_json_pathW[] = {'\\','w','i','n','e','v','u','l','k','a','n','.','j','s','o','n',0}; static const WCHAR vulkan_driversW[] = {'S','o','f','t','w','a','r','e','\\','K','h','r','o','n','o','s','\\', 'V','u','l','k','a','n','\\','D','r','i','v','e','r','s',0}; HRESULT WINAPI DllRegisterServer(void) { WCHAR json_path[MAX_PATH]; HRSRC rsrc; const char *data; DWORD datalen, written, zero = 0; HANDLE file; HKEY key; /* Create the JSON manifest and registry key to register this ICD with the official Vulkan loader. */ TRACE("\n"); rsrc = FindResourceW(hinstance, winevulkan_json_resW, (const WCHAR *)RT_RCDATA); data = LockResource(LoadResource(hinstance, rsrc)); datalen = SizeofResource(hinstance, rsrc); GetSystemDirectoryW(json_path, ARRAY_SIZE(json_path)); lstrcatW(json_path, winevulkan_json_pathW); file = CreateFileW(json_path, GENERIC_WRITE, 0, NULL, CREATE_ALWAYS, FILE_ATTRIBUTE_NORMAL, NULL); if (file == INVALID_HANDLE_VALUE) { ERR("Unable to create JSON manifest.\n"); return E_UNEXPECTED; } WriteFile(file, data, datalen, &written, NULL); CloseHandle(file); if (!RegCreateKeyExW(HKEY_LOCAL_MACHINE, vulkan_driversW, 0, NULL, 0, KEY_SET_VALUE, NULL, &key, NULL)) { RegSetValueExW(key, json_path, 0, REG_DWORD, (const BYTE *)&zero, sizeof(zero)); RegCloseKey(key); } return S_OK; } HRESULT WINAPI DllUnregisterServer(void) { WCHAR json_path[MAX_PATH]; HKEY key; /* Remove the JSON manifest and registry key */ TRACE("\n"); GetSystemDirectoryW(json_path, ARRAY_SIZE(json_path)); lstrcatW(json_path, winevulkan_json_pathW); DeleteFileW(json_path); if (RegOpenKeyExW(HKEY_LOCAL_MACHINE, vulkan_driversW, 0, KEY_SET_VALUE, &key) == ERROR_SUCCESS) { RegDeleteValueW(key, json_path); RegCloseKey(key); } return S_OK; }