Sweden-Number/dlls/wined3d/adapter_gl.c

5334 lines
247 KiB
C

/*
* Copyright 2002-2004 Jason Edmeades
* Copyright 2003-2004 Raphael Junqueira
* Copyright 2004 Christian Costa
* Copyright 2005 Oliver Stieber
* Copyright 2007-2008 Stefan Dösinger for CodeWeavers
* Copyright 2009-2011, 2018 Henri Verbeet for CodeWeavers
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free Software
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
*/
#include "config.h"
#include "wine/port.h"
#include <stdio.h>
#include "wined3d_private.h"
WINE_DEFAULT_DEBUG_CHANNEL(d3d);
WINE_DECLARE_DEBUG_CHANNEL(d3d_perf);
WINE_DECLARE_DEBUG_CHANNEL(winediag);
enum wined3d_gl_vendor
{
GL_VENDOR_UNKNOWN,
GL_VENDOR_APPLE,
GL_VENDOR_FGLRX,
GL_VENDOR_MESA,
GL_VENDOR_NVIDIA,
};
struct wined3d_extension_map
{
const char *extension_string;
enum wined3d_gl_extension extension;
};
static const struct wined3d_extension_map gl_extension_map[] =
{
/* APPLE */
{"GL_APPLE_fence", APPLE_FENCE },
{"GL_APPLE_float_pixels", APPLE_FLOAT_PIXELS },
{"GL_APPLE_flush_buffer_range", APPLE_FLUSH_BUFFER_RANGE },
{"GL_APPLE_ycbcr_422", APPLE_YCBCR_422 },
/* ARB */
{"GL_ARB_base_instance", ARB_BASE_INSTANCE },
{"GL_ARB_blend_func_extended", ARB_BLEND_FUNC_EXTENDED },
{"GL_ARB_buffer_storage", ARB_BUFFER_STORAGE },
{"GL_ARB_clear_buffer_object", ARB_CLEAR_BUFFER_OBJECT },
{"GL_ARB_clear_texture", ARB_CLEAR_TEXTURE },
{"GL_ARB_clip_control", ARB_CLIP_CONTROL },
{"GL_ARB_color_buffer_float", ARB_COLOR_BUFFER_FLOAT },
{"GL_ARB_compute_shader", ARB_COMPUTE_SHADER },
{"GL_ARB_conservative_depth", ARB_CONSERVATIVE_DEPTH },
{"GL_ARB_copy_buffer", ARB_COPY_BUFFER },
{"GL_ARB_copy_image", ARB_COPY_IMAGE },
{"GL_ARB_cull_distance", ARB_CULL_DISTANCE },
{"GL_ARB_debug_output", ARB_DEBUG_OUTPUT },
{"GL_ARB_depth_buffer_float", ARB_DEPTH_BUFFER_FLOAT },
{"GL_ARB_depth_clamp", ARB_DEPTH_CLAMP },
{"GL_ARB_depth_texture", ARB_DEPTH_TEXTURE },
{"GL_ARB_derivative_control", ARB_DERIVATIVE_CONTROL },
{"GL_ARB_draw_buffers", ARB_DRAW_BUFFERS },
{"GL_ARB_draw_elements_base_vertex", ARB_DRAW_ELEMENTS_BASE_VERTEX },
{"GL_ARB_draw_indirect", ARB_DRAW_INDIRECT },
{"GL_ARB_draw_instanced", ARB_DRAW_INSTANCED },
{"GL_ARB_ES2_compatibility", ARB_ES2_COMPATIBILITY },
{"GL_ARB_ES3_compatibility", ARB_ES3_COMPATIBILITY },
{"GL_ARB_explicit_attrib_location", ARB_EXPLICIT_ATTRIB_LOCATION },
{"GL_ARB_fragment_coord_conventions", ARB_FRAGMENT_COORD_CONVENTIONS},
{"GL_ARB_fragment_layer_viewport", ARB_FRAGMENT_LAYER_VIEWPORT },
{"GL_ARB_fragment_program", ARB_FRAGMENT_PROGRAM },
{"GL_ARB_fragment_shader", ARB_FRAGMENT_SHADER },
{"GL_ARB_framebuffer_no_attachments", ARB_FRAMEBUFFER_NO_ATTACHMENTS},
{"GL_ARB_framebuffer_object", ARB_FRAMEBUFFER_OBJECT },
{"GL_ARB_framebuffer_sRGB", ARB_FRAMEBUFFER_SRGB },
{"GL_ARB_geometry_shader4", ARB_GEOMETRY_SHADER4 },
{"GL_ARB_gpu_shader5", ARB_GPU_SHADER5 },
{"GL_ARB_half_float_pixel", ARB_HALF_FLOAT_PIXEL },
{"GL_ARB_half_float_vertex", ARB_HALF_FLOAT_VERTEX },
{"GL_ARB_instanced_arrays", ARB_INSTANCED_ARRAYS },
{"GL_ARB_internalformat_query", ARB_INTERNALFORMAT_QUERY },
{"GL_ARB_internalformat_query2", ARB_INTERNALFORMAT_QUERY2 },
{"GL_ARB_map_buffer_alignment", ARB_MAP_BUFFER_ALIGNMENT },
{"GL_ARB_map_buffer_range", ARB_MAP_BUFFER_RANGE },
{"GL_ARB_multisample", ARB_MULTISAMPLE },
{"GL_ARB_multitexture", ARB_MULTITEXTURE },
{"GL_ARB_occlusion_query", ARB_OCCLUSION_QUERY },
{"GL_ARB_pipeline_statistics_query", ARB_PIPELINE_STATISTICS_QUERY },
{"GL_ARB_pixel_buffer_object", ARB_PIXEL_BUFFER_OBJECT },
{"GL_ARB_point_parameters", ARB_POINT_PARAMETERS },
{"GL_ARB_point_sprite", ARB_POINT_SPRITE },
{"GL_ARB_polygon_offset_clamp", ARB_POLYGON_OFFSET_CLAMP },
{"GL_ARB_provoking_vertex", ARB_PROVOKING_VERTEX },
{"GL_ARB_query_buffer_object", ARB_QUERY_BUFFER_OBJECT },
{"GL_ARB_sample_shading", ARB_SAMPLE_SHADING },
{"GL_ARB_sampler_objects", ARB_SAMPLER_OBJECTS },
{"GL_ARB_seamless_cube_map", ARB_SEAMLESS_CUBE_MAP },
{"GL_ARB_shader_atomic_counters", ARB_SHADER_ATOMIC_COUNTERS },
{"GL_ARB_shader_bit_encoding", ARB_SHADER_BIT_ENCODING },
{"GL_ARB_shader_image_load_store", ARB_SHADER_IMAGE_LOAD_STORE },
{"GL_ARB_shader_image_size", ARB_SHADER_IMAGE_SIZE },
{"GL_ARB_shader_storage_buffer_object", ARB_SHADER_STORAGE_BUFFER_OBJECT},
{"GL_ARB_shader_texture_image_samples", ARB_SHADER_TEXTURE_IMAGE_SAMPLES},
{"GL_ARB_shader_texture_lod", ARB_SHADER_TEXTURE_LOD },
{"GL_ARB_shader_viewport_layer_array", ARB_SHADER_VIEWPORT_LAYER_ARRAY},
{"GL_ARB_shading_language_100", ARB_SHADING_LANGUAGE_100 },
{"GL_ARB_shading_language_420pack", ARB_SHADING_LANGUAGE_420PACK },
{"GL_ARB_shading_language_packing", ARB_SHADING_LANGUAGE_PACKING },
{"GL_ARB_shadow", ARB_SHADOW },
{"GL_ARB_stencil_texturing", ARB_STENCIL_TEXTURING },
{"GL_ARB_sync", ARB_SYNC },
{"GL_ARB_tessellation_shader", ARB_TESSELLATION_SHADER },
{"GL_ARB_texture_border_clamp", ARB_TEXTURE_BORDER_CLAMP },
{"GL_ARB_texture_buffer_object", ARB_TEXTURE_BUFFER_OBJECT },
{"GL_ARB_texture_buffer_range", ARB_TEXTURE_BUFFER_RANGE },
{"GL_ARB_texture_compression", ARB_TEXTURE_COMPRESSION },
{"GL_ARB_texture_compression_bptc", ARB_TEXTURE_COMPRESSION_BPTC },
{"GL_ARB_texture_compression_rgtc", ARB_TEXTURE_COMPRESSION_RGTC },
{"GL_ARB_texture_cube_map", ARB_TEXTURE_CUBE_MAP },
{"GL_ARB_texture_cube_map_array", ARB_TEXTURE_CUBE_MAP_ARRAY },
{"GL_ARB_texture_env_combine", ARB_TEXTURE_ENV_COMBINE },
{"GL_ARB_texture_env_dot3", ARB_TEXTURE_ENV_DOT3 },
{"GL_ARB_texture_filter_anisotropic", ARB_TEXTURE_FILTER_ANISOTROPIC},
{"GL_ARB_texture_float", ARB_TEXTURE_FLOAT },
{"GL_ARB_texture_gather", ARB_TEXTURE_GATHER },
{"GL_ARB_texture_mirrored_repeat", ARB_TEXTURE_MIRRORED_REPEAT },
{"GL_ARB_texture_mirror_clamp_to_edge", ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE},
{"GL_ARB_texture_multisample", ARB_TEXTURE_MULTISAMPLE },
{"GL_ARB_texture_non_power_of_two", ARB_TEXTURE_NON_POWER_OF_TWO },
{"GL_ARB_texture_query_levels", ARB_TEXTURE_QUERY_LEVELS },
{"GL_ARB_texture_rectangle", ARB_TEXTURE_RECTANGLE },
{"GL_ARB_texture_rg", ARB_TEXTURE_RG },
{"GL_ARB_texture_rgb10_a2ui", ARB_TEXTURE_RGB10_A2UI },
{"GL_ARB_texture_storage", ARB_TEXTURE_STORAGE },
{"GL_ARB_texture_storage_multisample", ARB_TEXTURE_STORAGE_MULTISAMPLE},
{"GL_ARB_texture_swizzle", ARB_TEXTURE_SWIZZLE },
{"GL_ARB_texture_view", ARB_TEXTURE_VIEW },
{"GL_ARB_timer_query", ARB_TIMER_QUERY },
{"GL_ARB_transform_feedback2", ARB_TRANSFORM_FEEDBACK2 },
{"GL_ARB_transform_feedback3", ARB_TRANSFORM_FEEDBACK3 },
{"GL_ARB_uniform_buffer_object", ARB_UNIFORM_BUFFER_OBJECT },
{"GL_ARB_vertex_array_bgra", ARB_VERTEX_ARRAY_BGRA },
{"GL_ARB_vertex_buffer_object", ARB_VERTEX_BUFFER_OBJECT },
{"GL_ARB_vertex_program", ARB_VERTEX_PROGRAM },
{"GL_ARB_vertex_shader", ARB_VERTEX_SHADER },
{"GL_ARB_vertex_type_2_10_10_10_rev", ARB_VERTEX_TYPE_2_10_10_10_REV},
{"GL_ARB_viewport_array", ARB_VIEWPORT_ARRAY },
/* ATI */
{"GL_ATI_fragment_shader", ATI_FRAGMENT_SHADER },
{"GL_ATI_separate_stencil", ATI_SEPARATE_STENCIL },
{"GL_ATI_texture_compression_3dc", ATI_TEXTURE_COMPRESSION_3DC },
{"GL_ATI_texture_env_combine3", ATI_TEXTURE_ENV_COMBINE3 },
{"GL_ATI_texture_mirror_once", ATI_TEXTURE_MIRROR_ONCE },
/* EXT */
{"GL_EXT_blend_color", EXT_BLEND_COLOR },
{"GL_EXT_blend_equation_separate", EXT_BLEND_EQUATION_SEPARATE },
{"GL_EXT_blend_func_separate", EXT_BLEND_FUNC_SEPARATE },
{"GL_EXT_blend_minmax", EXT_BLEND_MINMAX },
{"GL_EXT_blend_subtract", EXT_BLEND_SUBTRACT },
{"GL_EXT_depth_bounds_test", EXT_DEPTH_BOUNDS_TEST },
{"GL_EXT_draw_buffers2", EXT_DRAW_BUFFERS2 },
{"GL_EXT_fog_coord", EXT_FOG_COORD },
{"GL_EXT_framebuffer_blit", EXT_FRAMEBUFFER_BLIT },
{"GL_EXT_framebuffer_multisample", EXT_FRAMEBUFFER_MULTISAMPLE },
{"GL_EXT_framebuffer_object", EXT_FRAMEBUFFER_OBJECT },
{"GL_EXT_memory_object", EXT_MEMORY_OBJECT },
{"GL_EXT_gpu_program_parameters", EXT_GPU_PROGRAM_PARAMETERS },
{"GL_EXT_gpu_shader4", EXT_GPU_SHADER4 },
{"GL_EXT_packed_depth_stencil", EXT_PACKED_DEPTH_STENCIL },
{"GL_EXT_packed_float", EXT_PACKED_FLOAT },
{"GL_EXT_point_parameters", EXT_POINT_PARAMETERS },
{"GL_EXT_polygon_offset_clamp", ARB_POLYGON_OFFSET_CLAMP },
{"GL_EXT_provoking_vertex", EXT_PROVOKING_VERTEX },
{"GL_EXT_secondary_color", EXT_SECONDARY_COLOR },
{"GL_EXT_stencil_two_side", EXT_STENCIL_TWO_SIDE },
{"GL_EXT_stencil_wrap", EXT_STENCIL_WRAP },
{"GL_EXT_texture3D", EXT_TEXTURE3D },
{"GL_EXT_texture_array", EXT_TEXTURE_ARRAY },
{"GL_EXT_texture_compression_rgtc", EXT_TEXTURE_COMPRESSION_RGTC },
{"GL_EXT_texture_compression_s3tc", EXT_TEXTURE_COMPRESSION_S3TC },
{"GL_EXT_texture_env_combine", EXT_TEXTURE_ENV_COMBINE },
{"GL_EXT_texture_env_dot3", EXT_TEXTURE_ENV_DOT3 },
{"GL_EXT_texture_filter_anisotropic", ARB_TEXTURE_FILTER_ANISOTROPIC},
{"GL_EXT_texture_integer", EXT_TEXTURE_INTEGER },
{"GL_EXT_texture_lod_bias", EXT_TEXTURE_LOD_BIAS },
{"GL_EXT_texture_mirror_clamp", EXT_TEXTURE_MIRROR_CLAMP },
{"GL_EXT_texture_shadow_lod", EXT_TEXTURE_SHADOW_LOD },
{"GL_EXT_texture_shared_exponent", EXT_TEXTURE_SHARED_EXPONENT },
{"GL_EXT_texture_snorm", EXT_TEXTURE_SNORM },
{"GL_EXT_texture_sRGB", EXT_TEXTURE_SRGB },
{"GL_EXT_texture_sRGB_decode", EXT_TEXTURE_SRGB_DECODE },
{"GL_EXT_texture_swizzle", ARB_TEXTURE_SWIZZLE },
{"GL_EXT_vertex_array_bgra", ARB_VERTEX_ARRAY_BGRA },
/* NV */
{"GL_NV_fence", NV_FENCE },
{"GL_NV_fog_distance", NV_FOG_DISTANCE },
{"GL_NV_fragment_program", NV_FRAGMENT_PROGRAM },
{"GL_NV_fragment_program2", NV_FRAGMENT_PROGRAM2 },
{"GL_NV_fragment_program_option", NV_FRAGMENT_PROGRAM_OPTION },
{"GL_NV_half_float", NV_HALF_FLOAT },
{"GL_NV_light_max_exponent", NV_LIGHT_MAX_EXPONENT },
{"GL_NV_point_sprite", NV_POINT_SPRITE },
{"GL_NV_register_combiners", NV_REGISTER_COMBINERS },
{"GL_NV_register_combiners2", NV_REGISTER_COMBINERS2 },
{"GL_NV_texgen_reflection", NV_TEXGEN_REFLECTION },
{"GL_NV_texture_env_combine4", NV_TEXTURE_ENV_COMBINE4 },
{"GL_NV_texture_shader", NV_TEXTURE_SHADER },
{"GL_NV_texture_shader2", NV_TEXTURE_SHADER2 },
{"GL_NV_vertex_program", NV_VERTEX_PROGRAM },
{"GL_NV_vertex_program1_1", NV_VERTEX_PROGRAM1_1 },
{"GL_NV_vertex_program2", NV_VERTEX_PROGRAM2 },
{"GL_NV_vertex_program2_option", NV_VERTEX_PROGRAM2_OPTION },
{"GL_NV_vertex_program3", NV_VERTEX_PROGRAM3 },
};
static const struct wined3d_extension_map wgl_extension_map[] =
{
{"WGL_ARB_pixel_format", WGL_ARB_PIXEL_FORMAT },
{"WGL_EXT_swap_control", WGL_EXT_SWAP_CONTROL },
{"WGL_WINE_pixel_format_passthrough", WGL_WINE_PIXEL_FORMAT_PASSTHROUGH},
{"WGL_WINE_query_renderer", WGL_WINE_QUERY_RENDERER },
};
static void wined3d_caps_gl_ctx_destroy(const struct wined3d_caps_gl_ctx *ctx)
{
const struct wined3d_gl_info *gl_info = ctx->gl_info;
TRACE("Destroying caps GL context.\n");
/* Both glDeleteProgram and glDeleteBuffers silently ignore 0 IDs but
* this function might be called before the relevant function pointers
* in gl_info are initialized. */
if (ctx->test_program_id || ctx->test_vbo)
{
GL_EXTCALL(glDeleteProgram(ctx->test_program_id));
GL_EXTCALL(glDeleteBuffers(1, &ctx->test_vbo));
}
if (!wglMakeCurrent(NULL, NULL))
ERR("Failed to disable caps GL context.\n");
if (!wglDeleteContext(ctx->gl_ctx))
{
DWORD err = GetLastError();
ERR("wglDeleteContext(%p) failed, last error %#x.\n", ctx->gl_ctx, err);
}
wined3d_release_dc(ctx->wnd, ctx->dc);
DestroyWindow(ctx->wnd);
if (ctx->restore_gl_ctx && !wglMakeCurrent(ctx->restore_dc, ctx->restore_gl_ctx))
ERR("Failed to restore previous GL context.\n");
}
static BOOL wined3d_caps_gl_ctx_create_attribs(struct wined3d_caps_gl_ctx *caps_gl_ctx,
struct wined3d_gl_info *gl_info)
{
HGLRC new_ctx;
if (!(gl_info->p_wglCreateContextAttribsARB = (void *)wglGetProcAddress("wglCreateContextAttribsARB")))
return TRUE;
if (!(new_ctx = context_create_wgl_attribs(gl_info, caps_gl_ctx->dc, NULL)))
{
gl_info->p_wglCreateContextAttribsARB = NULL;
return FALSE;
}
if (!wglMakeCurrent(caps_gl_ctx->dc, new_ctx))
{
ERR("Failed to make new context current, last error %#x.\n", GetLastError());
if (!wglDeleteContext(new_ctx))
ERR("Failed to delete new context, last error %#x.\n", GetLastError());
gl_info->p_wglCreateContextAttribsARB = NULL;
return TRUE;
}
if (!wglDeleteContext(caps_gl_ctx->gl_ctx))
ERR("Failed to delete old context, last error %#x.\n", GetLastError());
caps_gl_ctx->gl_ctx = new_ctx;
return TRUE;
}
static BOOL wined3d_caps_gl_ctx_create(struct wined3d_adapter *adapter, struct wined3d_caps_gl_ctx *ctx)
{
PIXELFORMATDESCRIPTOR pfd;
int iPixelFormat;
TRACE("getting context...\n");
ctx->restore_dc = wglGetCurrentDC();
ctx->restore_gl_ctx = wglGetCurrentContext();
/* We need a fake window as a hdc retrieved using GetDC(0) can't be used for much GL purposes. */
ctx->wnd = CreateWindowA(WINED3D_OPENGL_WINDOW_CLASS_NAME, "WineD3D fake window",
WS_OVERLAPPEDWINDOW, 10, 10, 10, 10, NULL, NULL, NULL, NULL);
if (!ctx->wnd)
{
ERR("Failed to create a window.\n");
goto fail;
}
ctx->dc = GetDC(ctx->wnd);
if (!ctx->dc)
{
ERR("Failed to get a DC.\n");
goto fail;
}
/* PixelFormat selection */
ZeroMemory(&pfd, sizeof(pfd));
pfd.nSize = sizeof(pfd);
pfd.nVersion = 1;
pfd.dwFlags = PFD_SUPPORT_OPENGL | PFD_DOUBLEBUFFER | PFD_DRAW_TO_WINDOW; /* PFD_GENERIC_ACCELERATED */
pfd.iPixelType = PFD_TYPE_RGBA;
pfd.cColorBits = 32;
pfd.iLayerType = PFD_MAIN_PLANE;
if (!(iPixelFormat = ChoosePixelFormat(ctx->dc, &pfd)))
{
/* If this happens something is very wrong as ChoosePixelFormat barely fails. */
ERR("Failed to find a suitable pixel format.\n");
goto fail;
}
DescribePixelFormat(ctx->dc, iPixelFormat, sizeof(pfd), &pfd);
SetPixelFormat(ctx->dc, iPixelFormat, &pfd);
/* Create a GL context. */
if (!(ctx->gl_ctx = wglCreateContext(ctx->dc)))
{
WARN("Failed to create default context for capabilities initialization.\n");
goto fail;
}
/* Make it the current GL context. */
if (!wglMakeCurrent(ctx->dc, ctx->gl_ctx))
{
ERR("Failed to make caps GL context current.\n");
goto fail;
}
ctx->gl_info = &adapter->gl_info;
return TRUE;
fail:
if (ctx->gl_ctx) wglDeleteContext(ctx->gl_ctx);
ctx->gl_ctx = NULL;
if (ctx->dc) ReleaseDC(ctx->wnd, ctx->dc);
ctx->dc = NULL;
if (ctx->wnd) DestroyWindow(ctx->wnd);
ctx->wnd = NULL;
if (ctx->restore_gl_ctx && !wglMakeCurrent(ctx->restore_dc, ctx->restore_gl_ctx))
ERR("Failed to restore previous GL context.\n");
return FALSE;
}
/* Context activation is done by the caller. */
static BOOL test_arb_vs_offset_limit(const struct wined3d_gl_info *gl_info)
{
GLuint prog;
BOOL ret = FALSE;
static const char testcode[] =
"!!ARBvp1.0\n"
"PARAM C[66] = { program.env[0..65] };\n"
"ADDRESS A0;"
"PARAM zero = {0.0, 0.0, 0.0, 0.0};\n"
"ARL A0.x, zero.x;\n"
"MOV result.position, C[A0.x + 65];\n"
"END\n";
while (gl_info->gl_ops.gl.p_glGetError());
GL_EXTCALL(glGenProgramsARB(1, &prog));
if(!prog) {
ERR("Failed to create an ARB offset limit test program\n");
}
GL_EXTCALL(glBindProgramARB(GL_VERTEX_PROGRAM_ARB, prog));
GL_EXTCALL(glProgramStringARB(GL_VERTEX_PROGRAM_ARB, GL_PROGRAM_FORMAT_ASCII_ARB,
strlen(testcode), testcode));
if (gl_info->gl_ops.gl.p_glGetError())
{
TRACE("OpenGL implementation does not allow indirect addressing offsets > 63\n");
TRACE("error: %s\n", debugstr_a((const char *)gl_info->gl_ops.gl.p_glGetString(GL_PROGRAM_ERROR_STRING_ARB)));
ret = TRUE;
} else TRACE("OpenGL implementation allows offsets > 63\n");
GL_EXTCALL(glBindProgramARB(GL_VERTEX_PROGRAM_ARB, 0));
GL_EXTCALL(glDeleteProgramsARB(1, &prog));
checkGLcall("ARB vp offset limit test cleanup");
return ret;
}
static BOOL match_amd_r300_to_500(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
return card_vendor == HW_VENDOR_AMD
&& (device == CARD_AMD_RADEON_9500
|| device == CARD_AMD_RADEON_X700
|| device == CARD_AMD_RADEON_X1600);
}
static BOOL match_geforce5(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
return card_vendor == HW_VENDOR_NVIDIA
&& (device == CARD_NVIDIA_GEFORCEFX_5200
|| device == CARD_NVIDIA_GEFORCEFX_5600
|| device == CARD_NVIDIA_GEFORCEFX_5800);
}
static BOOL match_apple(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
/* MacOS has various specialities in the extensions it advertises. Some
* have to be loaded from the OpenGL 1.2+ core, while other extensions are
* advertised, but software emulated. So try to detect the Apple OpenGL
* implementation to apply some extension fixups afterwards.
*
* Detecting this isn't really easy. The vendor string doesn't mention
* Apple. Compile-time checks aren't sufficient either because a Linux
* binary may display on a macOS X server via remote X11. So try to detect
* the OpenGL implementation by looking at certain Apple extensions. Some
* extensions like client storage might be supported on other
* implementations too, but GL_APPLE_flush_render is specific to the
* macOS X window management, and GL_APPLE_ycbcr_422 is QuickTime
* specific. So the chance that other implementations support them is
* rather small since Win32 QuickTime uses DirectDraw, not OpenGL.
*
* This test has been moved into wined3d_guess_gl_vendor(). */
return gl_vendor == GL_VENDOR_APPLE;
}
/* Context activation is done by the caller. */
static void test_pbo_functionality(struct wined3d_gl_info *gl_info)
{
/* Some OpenGL implementations, namely Apple's Geforce 8 driver, advertises PBOs,
* but glTexSubImage from a PBO fails miserably, with the first line repeated over
* all the texture. This function detects this bug by its symptom and disables PBOs
* if the test fails.
*
* The test uploads a 4x4 texture via the PBO in the "native" format GL_BGRA,
* GL_UNSIGNED_INT_8_8_8_8_REV. This format triggers the bug, and it is what we use
* for D3DFMT_A8R8G8B8. Then the texture is read back without any PBO and the data
* read back is compared to the original. If they are equal PBOs are assumed to work,
* otherwise the PBO extension is disabled. */
GLuint texture, pbo;
static const unsigned int pattern[] =
{
0x00000000, 0x000000ff, 0x0000ff00, 0x40ff0000,
0x80ffffff, 0x40ffff00, 0x00ff00ff, 0x0000ffff,
0x00ffff00, 0x00ff00ff, 0x0000ffff, 0x000000ff,
0x80ff00ff, 0x0000ffff, 0x00ff00ff, 0x40ff00ff
};
unsigned int check[ARRAY_SIZE(pattern)];
/* No PBO -> No point in testing them. */
if (!gl_info->supported[ARB_PIXEL_BUFFER_OBJECT]) return;
while (gl_info->gl_ops.gl.p_glGetError());
gl_info->gl_ops.gl.p_glGenTextures(1, &texture);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, texture);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAX_LEVEL, 0);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA8, 4, 4, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, 0);
checkGLcall("Specifying the PBO test texture");
GL_EXTCALL(glGenBuffers(1, &pbo));
GL_EXTCALL(glBindBuffer(GL_PIXEL_UNPACK_BUFFER, pbo));
GL_EXTCALL(glBufferData(GL_PIXEL_UNPACK_BUFFER, sizeof(pattern), pattern, GL_STREAM_DRAW));
checkGLcall("Specifying the PBO test pbo");
gl_info->gl_ops.gl.p_glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 4, 4, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, NULL);
checkGLcall("Loading the PBO test texture");
GL_EXTCALL(glBindBuffer(GL_PIXEL_UNPACK_BUFFER, 0));
gl_info->gl_ops.gl.p_glFinish(); /* just to be sure */
memset(check, 0, sizeof(check));
gl_info->gl_ops.gl.p_glGetTexImage(GL_TEXTURE_2D, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, check);
checkGLcall("Reading back the PBO test texture");
gl_info->gl_ops.gl.p_glDeleteTextures(1, &texture);
GL_EXTCALL(glDeleteBuffers(1, &pbo));
checkGLcall("PBO test cleanup");
if (memcmp(check, pattern, sizeof(check)))
{
WARN_(d3d_perf)("PBO test failed, read back data doesn't match original.\n"
"Disabling PBOs. This may result in slower performance.\n");
gl_info->supported[ARB_PIXEL_BUFFER_OBJECT] = FALSE;
}
else
{
TRACE("PBO test successful.\n");
}
}
static BOOL match_apple_intel(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
return (card_vendor == HW_VENDOR_INTEL) && (gl_vendor == GL_VENDOR_APPLE);
}
static BOOL match_apple_nonr500ati(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
if (gl_vendor != GL_VENDOR_APPLE) return FALSE;
if (card_vendor != HW_VENDOR_AMD) return FALSE;
if (device == CARD_AMD_RADEON_X1600) return FALSE;
return TRUE;
}
static BOOL match_dx10_capable(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
/* Direct3D 9 cards support 40 single float varyings in hardware, most
* drivers report 32. ATI misreports 44 varyings. So assume that if we
* have more than 44 varyings we have a Direct3D 10+ card. This detection
* is for the gl_ClipPos varying quirk. If a Direct3D 9 card really
* supports more than 44 varyings and we subtract one in Direct3D 9
* shaders it's not going to hurt us because the Direct3D 9 limit is
* hardcoded.
*
* Direct3D 10 cards usually have 64 varyings. */
return gl_info->limits.glsl_varyings > 44;
}
static BOOL match_not_dx10_capable(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
return !match_dx10_capable(gl_info, ctx, gl_renderer, gl_vendor, card_vendor, device);
}
/* A GL context is provided by the caller */
static BOOL match_allows_spec_alpha(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
GLenum error;
DWORD data[16];
if (!gl_info->supported[EXT_SECONDARY_COLOR] || !gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
return FALSE;
while (gl_info->gl_ops.gl.p_glGetError());
GL_EXTCALL(glSecondaryColorPointerEXT)(4, GL_UNSIGNED_BYTE, 4, data);
error = gl_info->gl_ops.gl.p_glGetError();
if (error == GL_NO_ERROR)
{
TRACE("GL Implementation accepts 4 component specular color pointers\n");
return TRUE;
}
else
{
TRACE("GL implementation does not accept 4 component specular colors, error %s\n",
debug_glerror(error));
return FALSE;
}
}
/* A GL context is provided by the caller */
static BOOL match_broken_nv_clip(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
GLuint prog;
BOOL ret = FALSE;
GLint pos;
static const char testcode[] =
"!!ARBvp1.0\n"
"OPTION NV_vertex_program2;\n"
"MOV result.clip[0], 0.0;\n"
"MOV result.position, 0.0;\n"
"END\n";
if (!gl_info->supported[NV_VERTEX_PROGRAM2_OPTION]) return FALSE;
while (gl_info->gl_ops.gl.p_glGetError());
GL_EXTCALL(glGenProgramsARB(1, &prog));
if(!prog)
{
ERR("Failed to create the NVvp clip test program\n");
return FALSE;
}
GL_EXTCALL(glBindProgramARB(GL_VERTEX_PROGRAM_ARB, prog));
GL_EXTCALL(glProgramStringARB(GL_VERTEX_PROGRAM_ARB, GL_PROGRAM_FORMAT_ASCII_ARB,
strlen(testcode), testcode));
gl_info->gl_ops.gl.p_glGetIntegerv(GL_PROGRAM_ERROR_POSITION_ARB, &pos);
if(pos != -1)
{
WARN("GL_NV_vertex_program2_option result.clip[] test failed\n");
TRACE("error: %s\n", debugstr_a((const char *)gl_info->gl_ops.gl.p_glGetString(GL_PROGRAM_ERROR_STRING_ARB)));
ret = TRUE;
while (gl_info->gl_ops.gl.p_glGetError());
}
else TRACE("GL_NV_vertex_program2_option result.clip[] test passed\n");
GL_EXTCALL(glBindProgramARB(GL_VERTEX_PROGRAM_ARB, 0));
GL_EXTCALL(glDeleteProgramsARB(1, &prog));
checkGLcall("GL_NV_vertex_program2_option result.clip[] test cleanup");
return ret;
}
/* Context activation is done by the caller. */
static BOOL match_fbo_tex_update(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
char data[4 * 4 * 4];
GLuint tex, fbo;
GLenum status;
if (wined3d_settings.offscreen_rendering_mode != ORM_FBO) return FALSE;
memset(data, 0xcc, sizeof(data));
gl_info->gl_ops.gl.p_glGenTextures(1, &tex);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, tex);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA8, 4, 4, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, NULL);
checkGLcall("glTexImage2D");
gl_info->fbo_ops.glGenFramebuffers(1, &fbo);
gl_info->fbo_ops.glBindFramebuffer(GL_FRAMEBUFFER, fbo);
gl_info->fbo_ops.glFramebufferTexture2D(GL_FRAMEBUFFER, GL_COLOR_ATTACHMENT0, GL_TEXTURE_2D, tex, 0);
checkGLcall("glFramebufferTexture2D");
status = gl_info->fbo_ops.glCheckFramebufferStatus(GL_FRAMEBUFFER);
if (status != GL_FRAMEBUFFER_COMPLETE) ERR("FBO status %#x\n", status);
checkGLcall("glCheckFramebufferStatus");
memset(data, 0x11, sizeof(data));
gl_info->gl_ops.gl.p_glTexSubImage2D(GL_TEXTURE_2D, 0, 0, 0, 4, 4, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, data);
checkGLcall("glTexSubImage2D");
gl_info->gl_ops.gl.p_glClearColor(0.996f, 0.729f, 0.745f, 0.792f);
gl_info->gl_ops.gl.p_glClear(GL_COLOR_BUFFER_BIT);
checkGLcall("glClear");
gl_info->gl_ops.gl.p_glGetTexImage(GL_TEXTURE_2D, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, data);
checkGLcall("glGetTexImage");
gl_info->fbo_ops.glFramebufferTexture2D(GL_FRAMEBUFFER, GL_COLOR_ATTACHMENT0, GL_TEXTURE_2D, 0, 0);
gl_info->fbo_ops.glBindFramebuffer(GL_FRAMEBUFFER, 0);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, 0);
checkGLcall("glBindTexture");
gl_info->fbo_ops.glDeleteFramebuffers(1, &fbo);
gl_info->gl_ops.gl.p_glDeleteTextures(1, &tex);
checkGLcall("glDeleteTextures");
return *(DWORD *)data == 0x11111111;
}
/* Context activation is done by the caller. */
static BOOL match_broken_rgba16(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
/* GL_RGBA16 uses GL_RGBA8 internally on Geforce 7 and older cards.
* This leads to graphical bugs in Half Life 2 and Unreal engine games. */
GLuint tex;
GLint size;
gl_info->gl_ops.gl.p_glGenTextures(1, &tex);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, tex);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA16, 4, 4, 0, GL_RGBA, GL_UNSIGNED_SHORT, NULL);
checkGLcall("glTexImage2D");
gl_info->gl_ops.gl.p_glGetTexLevelParameteriv(GL_TEXTURE_2D, 0, GL_TEXTURE_RED_SIZE, &size);
checkGLcall("glGetTexLevelParameteriv");
TRACE("Real color depth is %d\n", size);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, 0);
checkGLcall("glBindTexture");
gl_info->gl_ops.gl.p_glDeleteTextures(1, &tex);
checkGLcall("glDeleteTextures");
return size < 16;
}
static BOOL match_fglrx(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
return gl_vendor == GL_VENDOR_FGLRX;
}
static BOOL match_r200(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
if (card_vendor != HW_VENDOR_AMD) return FALSE;
if (device == CARD_AMD_RADEON_8500) return TRUE;
return FALSE;
}
static BOOL match_broken_arb_fog(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
DWORD data[4];
GLuint tex, fbo;
GLenum status;
float color[4] = {0.0f, 1.0f, 0.0f, 0.0f};
GLuint prog;
GLint err_pos;
static const char program_code[] =
"!!ARBfp1.0\n"
"OPTION ARB_fog_linear;\n"
"MOV result.color, {1.0, 0.0, 0.0, 0.0};\n"
"END\n";
if (wined3d_settings.offscreen_rendering_mode != ORM_FBO)
return FALSE;
if (!gl_info->supported[ARB_FRAGMENT_PROGRAM])
return FALSE;
if (!gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
return FALSE;
gl_info->gl_ops.gl.p_glGenTextures(1, &tex);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, tex);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB8, 4, 1, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, NULL);
checkGLcall("glTexImage2D");
gl_info->fbo_ops.glGenFramebuffers(1, &fbo);
gl_info->fbo_ops.glBindFramebuffer(GL_FRAMEBUFFER, fbo);
gl_info->fbo_ops.glFramebufferTexture2D(GL_FRAMEBUFFER, GL_COLOR_ATTACHMENT0, GL_TEXTURE_2D, tex, 0);
checkGLcall("glFramebufferTexture2D");
status = gl_info->fbo_ops.glCheckFramebufferStatus(GL_FRAMEBUFFER);
if (status != GL_FRAMEBUFFER_COMPLETE) ERR("FBO status %#x\n", status);
checkGLcall("glCheckFramebufferStatus");
gl_info->gl_ops.gl.p_glClearColor(0.0f, 0.0f, 1.0f, 0.0f);
gl_info->gl_ops.gl.p_glClear(GL_COLOR_BUFFER_BIT);
checkGLcall("glClear");
gl_info->gl_ops.gl.p_glViewport(0, 0, 4, 1);
checkGLcall("glViewport");
gl_info->gl_ops.gl.p_glEnable(GL_FOG);
gl_info->gl_ops.gl.p_glFogf(GL_FOG_START, 0.5f);
gl_info->gl_ops.gl.p_glFogf(GL_FOG_END, 0.5f);
gl_info->gl_ops.gl.p_glFogi(GL_FOG_MODE, GL_LINEAR);
gl_info->gl_ops.gl.p_glHint(GL_FOG_HINT, GL_NICEST);
gl_info->gl_ops.gl.p_glFogfv(GL_FOG_COLOR, color);
checkGLcall("fog setup");
GL_EXTCALL(glGenProgramsARB(1, &prog));
GL_EXTCALL(glBindProgramARB(GL_FRAGMENT_PROGRAM_ARB, prog));
GL_EXTCALL(glProgramStringARB(GL_FRAGMENT_PROGRAM_ARB, GL_PROGRAM_FORMAT_ASCII_ARB,
strlen(program_code), program_code));
gl_info->gl_ops.gl.p_glEnable(GL_FRAGMENT_PROGRAM_ARB);
checkGLcall("Test fragment program setup");
gl_info->gl_ops.gl.p_glGetIntegerv(GL_PROGRAM_ERROR_POSITION_ARB, &err_pos);
if (err_pos != -1)
{
const char *error_str;
error_str = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_PROGRAM_ERROR_STRING_ARB);
FIXME("Fog test program error at position %d: %s\n\n", err_pos, debugstr_a(error_str));
}
gl_info->gl_ops.gl.p_glBegin(GL_TRIANGLE_STRIP);
gl_info->gl_ops.gl.p_glVertex3f(-1.0f, -1.0f, 0.0f);
gl_info->gl_ops.gl.p_glVertex3f( 1.0f, -1.0f, 1.0f);
gl_info->gl_ops.gl.p_glVertex3f(-1.0f, 1.0f, 0.0f);
gl_info->gl_ops.gl.p_glVertex3f( 1.0f, 1.0f, 1.0f);
gl_info->gl_ops.gl.p_glEnd();
checkGLcall("ARBfp fog test draw");
gl_info->gl_ops.gl.p_glGetTexImage(GL_TEXTURE_2D, 0, GL_BGRA, GL_UNSIGNED_INT_8_8_8_8_REV, data);
checkGLcall("glGetTexImage");
data[0] &= 0x00ffffff;
data[1] &= 0x00ffffff;
data[2] &= 0x00ffffff;
data[3] &= 0x00ffffff;
gl_info->fbo_ops.glBindFramebuffer(GL_FRAMEBUFFER, 0);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, 0);
gl_info->fbo_ops.glDeleteFramebuffers(1, &fbo);
gl_info->gl_ops.gl.p_glDeleteTextures(1, &tex);
gl_info->gl_ops.gl.p_glDisable(GL_FOG);
GL_EXTCALL(glBindProgramARB(GL_FRAGMENT_PROGRAM_ARB, 0));
gl_info->gl_ops.gl.p_glDisable(GL_FRAGMENT_PROGRAM_ARB);
GL_EXTCALL(glDeleteProgramsARB(1, &prog));
checkGLcall("ARBfp fog test teardown");
TRACE("Fog test data: %08x %08x %08x %08x\n", data[0], data[1], data[2], data[3]);
return data[0] != 0x00ff0000 || data[3] != 0x0000ff00;
}
static BOOL match_broken_viewport_subpixel_bits(const struct wined3d_gl_info *gl_info,
struct wined3d_caps_gl_ctx *ctx, const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
if (!gl_info->supported[ARB_VIEWPORT_ARRAY])
return FALSE;
if (wined3d_settings.offscreen_rendering_mode != ORM_FBO)
return FALSE;
return !wined3d_caps_gl_ctx_test_viewport_subpixel_bits(ctx);
}
static BOOL match_no_independent_bit_depths(const struct wined3d_gl_info *gl_info,
struct wined3d_caps_gl_ctx *ctx, const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
GLuint tex[2], fbo;
GLenum status;
/* ARB_framebuffer_object allows implementation-dependent internal format
* restrictions. The EXT extension explicitly calls out an error in the
* relevant case. */
if (!gl_info->supported[ARB_FRAMEBUFFER_OBJECT])
return TRUE;
if (wined3d_settings.offscreen_rendering_mode != ORM_FBO)
return TRUE;
gl_info->gl_ops.gl.p_glGenTextures(2, tex);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, tex[0]);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGBA8, 4, 1, 0, GL_RGBA, GL_UNSIGNED_INT_8_8_8_8, NULL);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, tex[1]);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_NEAREST);
gl_info->gl_ops.gl.p_glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB5, 4, 1, 0, GL_RGB, GL_UNSIGNED_SHORT_5_6_5, NULL);
gl_info->gl_ops.gl.p_glBindTexture(GL_TEXTURE_2D, 0);
gl_info->fbo_ops.glGenFramebuffers(1, &fbo);
gl_info->fbo_ops.glBindFramebuffer(GL_DRAW_FRAMEBUFFER, fbo);
gl_info->fbo_ops.glFramebufferTexture2D(GL_DRAW_FRAMEBUFFER, GL_COLOR_ATTACHMENT0, GL_TEXTURE_2D, tex[0], 0);
gl_info->fbo_ops.glFramebufferTexture2D(GL_DRAW_FRAMEBUFFER, GL_COLOR_ATTACHMENT1, GL_TEXTURE_2D, tex[1], 0);
status = gl_info->fbo_ops.glCheckFramebufferStatus(GL_DRAW_FRAMEBUFFER);
gl_info->fbo_ops.glBindFramebuffer(GL_DRAW_FRAMEBUFFER, 0);
gl_info->fbo_ops.glDeleteFramebuffers(1, &fbo);
gl_info->gl_ops.gl.p_glDeleteTextures(2, tex);
checkGLcall("testing multiple framebuffer attachments with different bit depths");
return status != GL_FRAMEBUFFER_COMPLETE;
}
static void quirk_apple_glsl_constants(struct wined3d_gl_info *gl_info)
{
/* MacOS needs uniforms for relative addressing offsets. This can
* accumulate to quite a few uniforms. Beyond that the general uniform
* isn't optimal, so reserve a number of uniforms. 12 vec4's should allow
* 48 different offsets or other helper immediate values. */
TRACE("Reserving 12 GLSL constants for compiler private use.\n");
gl_info->reserved_glsl_constants = max(gl_info->reserved_glsl_constants, 12);
}
static void quirk_amd_dx9(struct wined3d_gl_info *gl_info)
{
/* MacOS advertises GL_ARB_texture_non_power_of_two on ATI r500 and
* earlier cards, although these cards only support
* GL_ARB_texture_rectangle (D3DPTEXTURECAPS_NONPOW2CONDITIONAL).
*
* If real NP2 textures are used, the driver falls back to software. We
* could just disable the extension and use GL_ARB_texture_rectangle
* instead, but texture_rectangle is inconvenient due to the
* non-normalised texture coordinates. Thus set an internal extension
* flag, GL_WINE_normalized_texrect, which signals the code that it can
* use non-power-of-two textures as per GL_ARB_texture_non_power_of_two,
* but has to stick to the texture_rectangle limits.
*
* Fglrx doesn't advertise GL_ARB_texture_non_power_of_two, but it
* advertises OpenGL 2.0, which has this extension promoted to core. The
* extension loading code sets this extension supported due to that, so
* this code works on fglrx as well. */
if (gl_info->supported[ARB_TEXTURE_NON_POWER_OF_TWO])
{
TRACE("GL_ARB_texture_non_power_of_two advertised on R500 or earlier card, removing.\n");
gl_info->supported[ARB_TEXTURE_NON_POWER_OF_TWO] = FALSE;
gl_info->supported[WINED3D_GL_NORMALIZED_TEXRECT] = TRUE;
}
}
static void quirk_no_np2(struct wined3d_gl_info *gl_info)
{
/* The NVIDIA GeForce FX series reports OpenGL 2.0 capabilities with the
* latest drivers versions, but doesn't explicitly advertise the
* ARB_tex_npot extension in the OpenGL extension string. This usually
* means that ARB_tex_npot is supported in hardware as long as the
* application is staying within the limits enforced by the
* ARB_texture_rectangle extension. This however is not true for the
* FX series, which instantly falls back to a slower software path as
* soon as ARB_tex_npot is used. We therefore completely remove
* ARB_tex_npot from the list of supported extensions.
*
* Note that WINE_normalized_texrect can't be used in this case because
* internally it uses ARB_tex_npot, triggering the software fallback.
* There is not much we can do here apart from disabling the
* software-emulated extension and re-enable ARB_tex_rect (which was
* previously disabled in wined3d_adapter_init_gl_caps).
*
* This fixup removes performance problems on both the FX 5900 and
* FX 5700 (e.g. for framebuffer post-processing effects in the game
* "Max Payne 2"). The behaviour can be verified through a simple test
* app attached in bugreport #14724. */
TRACE("GL_ARB_texture_non_power_of_two advertised through OpenGL 2.0 on NV FX card, removing.\n");
gl_info->supported[ARB_TEXTURE_NON_POWER_OF_TWO] = FALSE;
gl_info->supported[ARB_TEXTURE_RECTANGLE] = TRUE;
}
static void quirk_texcoord_w(struct wined3d_gl_info *gl_info)
{
/* The Intel GPUs on macOS set the .w register of texcoords to 0.0 by
* default, which causes problems with fixed-function fragment processing.
* Ideally this flag should be detected with a test shader and OpenGL
* feedback mode, but some OpenGL implementations (macOS ATI at least,
* probably all macOS ones) do not like vertex shaders in feedback mode
* and return an error, even though it should be valid according to the
* spec.
*
* We don't want to enable this on all cards, as it adds an extra
* instruction per texcoord used. This makes the shader slower and eats
* instruction slots which should be available to the Direct3D
* application.
*
* ATI Radeon HD 2xxx cards on macOS have the issue. Instead of checking
* for the buggy cards, blacklist all Radeon cards on macOS and whitelist
* the good ones. That way we're prepared for the future. If this
* workaround is activated on cards that do not need it, it won't break
* things, just affect performance negatively. */
TRACE("Enabling vertex texture coord fixes in vertex shaders.\n");
gl_info->quirks |= WINED3D_QUIRK_SET_TEXCOORD_W;
}
static void quirk_clip_varying(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_GLSL_CLIP_VARYING;
}
static void quirk_allows_specular_alpha(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_ALLOWS_SPECULAR_ALPHA;
}
static void quirk_disable_nvvp_clip(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_NV_CLIP_BROKEN;
}
static void quirk_fbo_tex_update(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_FBO_TEX_UPDATE;
}
static void quirk_broken_rgba16(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_BROKEN_RGBA16;
}
static void quirk_infolog_spam(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_INFO_LOG_SPAM;
}
static void quirk_limited_tex_filtering(struct wined3d_gl_info *gl_info)
{
/* NVIDIA GeForce 6xxx and 7xxx support accelerated VTF only on a few
* selected texture formats. They are apparently the only Direct3D 9 class
* GPUs supporting VTF. Also, Direct3D 9-era GPUs are somewhat limited
* with float texture filtering and blending. */
gl_info->quirks |= WINED3D_QUIRK_LIMITED_TEX_FILTERING;
}
static void quirk_r200_constants(struct wined3d_gl_info *gl_info)
{
/* The Mesa r200 driver (and there is no other driver for this GPU Wine
* would run on) loads some fog parameters (start, end, exponent, but not
* the colour) into the program.
*
* Apparently the fog hardware is only able to handle linear fog with a
* range of 0.0;1.0, and it is the responsibility of the vertex pipeline
* to handle non-linear fog and linear fog with start and end other than
* 0.0 and 1.0. */
TRACE("Reserving 1 ARB constant for compiler private use.\n");
gl_info->reserved_arb_constants = max(gl_info->reserved_arb_constants, 1);
}
static void quirk_broken_arb_fog(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_BROKEN_ARB_FOG;
}
static void quirk_broken_viewport_subpixel_bits(struct wined3d_gl_info *gl_info)
{
if (gl_info->supported[ARB_CLIP_CONTROL])
{
TRACE("Disabling ARB_clip_control.\n");
gl_info->supported[ARB_CLIP_CONTROL] = FALSE;
}
}
static void quirk_no_independent_bit_depths(struct wined3d_gl_info *gl_info)
{
gl_info->quirks |= WINED3D_QUIRK_NO_INDEPENDENT_BIT_DEPTHS;
}
static const struct wined3d_gpu_description *query_gpu_description(const struct wined3d_gl_info *gl_info,
UINT64 *vram_bytes)
{
const struct wined3d_gpu_description *gpu_description = NULL, *gpu_description_override;
enum wined3d_pci_vendor vendor = PCI_VENDOR_NONE;
enum wined3d_pci_device device = PCI_DEVICE_NONE;
GLuint value;
if (gl_info->supported[WGL_WINE_QUERY_RENDERER])
{
if (GL_EXTCALL(wglQueryCurrentRendererIntegerWINE(WGL_RENDERER_VENDOR_ID_WINE, &value)))
vendor = value;
if (GL_EXTCALL(wglQueryCurrentRendererIntegerWINE(WGL_RENDERER_DEVICE_ID_WINE, &value)))
device = value;
if (GL_EXTCALL(wglQueryCurrentRendererIntegerWINE(WGL_RENDERER_VIDEO_MEMORY_WINE, &value)))
*vram_bytes = (UINT64)value * 1024 * 1024;
TRACE("Card reports vendor PCI ID 0x%04x, device PCI ID 0x%04x, 0x%s bytes of video memory.\n",
vendor, device, wine_dbgstr_longlong(*vram_bytes));
gpu_description = wined3d_get_gpu_description(vendor, device);
}
if ((gpu_description_override = wined3d_get_user_override_gpu_description(vendor, device)))
gpu_description = gpu_description_override;
return gpu_description;
}
/* Context activation is done by the caller. */
static void fixup_extensions(struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device)
{
unsigned int i;
static const struct driver_quirk
{
BOOL (*match)(const struct wined3d_gl_info *gl_info, struct wined3d_caps_gl_ctx *ctx,
const char *gl_renderer, enum wined3d_gl_vendor gl_vendor,
enum wined3d_pci_vendor card_vendor, enum wined3d_pci_device device);
void (*apply)(struct wined3d_gl_info *gl_info);
const char *description;
}
quirk_table[] =
{
{
match_amd_r300_to_500,
quirk_amd_dx9,
"AMD normalised texrect quirk"
},
{
match_apple,
quirk_apple_glsl_constants,
"Apple GLSL uniform override"
},
{
match_geforce5,
quirk_no_np2,
"Geforce 5 NP2 disable"
},
{
match_apple_intel,
quirk_texcoord_w,
"Init texcoord .w for Apple Intel GPU driver"
},
{
match_apple_nonr500ati,
quirk_texcoord_w,
"Init texcoord .w for Apple ATI >= r600 GPU driver"
},
{
match_dx10_capable,
quirk_clip_varying,
"Reserved varying for gl_ClipPos"
},
{
/* GL_EXT_secondary_color does not allow 4 component secondary
* colours, but most OpenGL implementations accept it. Apple's
* is the only OpenGL implementation known to reject it.
*
* If we can pass 4-component specular colours, do it, because
* (a) we don't have to screw around with the data, and (b) the
* Direct3D fixed-function vertex pipeline passes specular alpha
* to the pixel shader if any is used. Otherwise the specular
* alpha is used to pass the fog coordinate, which we pass to
* OpenGL via GL_EXT_fog_coord. */
match_allows_spec_alpha,
quirk_allows_specular_alpha,
"Allow specular alpha quirk"
},
{
match_broken_nv_clip,
quirk_disable_nvvp_clip,
"Apple NV_vertex_program clip bug quirk"
},
{
match_fbo_tex_update,
quirk_fbo_tex_update,
"FBO rebind for attachment updates"
},
{
match_broken_rgba16,
quirk_broken_rgba16,
"True RGBA16 is not available"
},
{
match_fglrx,
quirk_infolog_spam,
"Not printing GLSL infolog"
},
{
match_not_dx10_capable,
quirk_limited_tex_filtering,
"Texture filtering, blending and VTF support is limited"
},
{
match_r200,
quirk_r200_constants,
"r200 vertex shader constants"
},
{
match_broken_arb_fog,
quirk_broken_arb_fog,
"ARBfp fogstart == fogend workaround"
},
{
match_broken_viewport_subpixel_bits,
quirk_broken_viewport_subpixel_bits,
"NVIDIA viewport subpixel bits bug"
},
{
match_no_independent_bit_depths,
quirk_no_independent_bit_depths,
"No support for MRT with independent bit depths"
},
};
for (i = 0; i < ARRAY_SIZE(quirk_table); ++i)
{
if (!quirk_table[i].match(gl_info, ctx, gl_renderer, gl_vendor, card_vendor, device)) continue;
TRACE("Applying driver quirk \"%s\".\n", quirk_table[i].description);
quirk_table[i].apply(gl_info);
}
/* Find out if PBOs work as they are supposed to. */
test_pbo_functionality(gl_info);
}
static DWORD wined3d_parse_gl_version(const char *gl_version)
{
const char *ptr = gl_version;
int major, minor;
major = atoi(ptr);
if (major <= 0)
ERR("Invalid OpenGL major version %d.\n", major);
while (isdigit(*ptr)) ++ptr;
if (*ptr++ != '.')
ERR("Invalid OpenGL version string %s.\n", debugstr_a(gl_version));
minor = atoi(ptr);
TRACE("Found OpenGL version %d.%d.\n", major, minor);
return MAKEDWORD_VERSION(major, minor);
}
static enum wined3d_gl_vendor wined3d_guess_gl_vendor(const struct wined3d_gl_info *gl_info,
const char *gl_vendor_string, const char *gl_renderer, const char *gl_version)
{
/* MacOS has various specialities in the extensions it advertises. Some have to be loaded from
* the opengl 1.2+ core, while other extensions are advertised, but software emulated. So try to
* detect the Apple OpenGL implementation to apply some extension fixups afterwards.
*
* Detecting this isn't really easy. The vendor string doesn't mention Apple. Compile-time checks
* aren't sufficient either because a Linux binary may display on a macos X server via remote X11.
* So try to detect the GL implementation by looking at certain Apple extensions. Some extensions
* like client storage might be supported on other implementations too, but GL_APPLE_flush_render
* is specific to the Mac OS X window management, and GL_APPLE_ycbcr_422 is QuickTime specific. So
* the chance that other implementations support them is rather small since Win32 QuickTime uses
* DirectDraw, not OpenGL. */
if (gl_info->supported[APPLE_FENCE] && gl_info->supported[APPLE_YCBCR_422])
return GL_VENDOR_APPLE;
if (strstr(gl_vendor_string, "NVIDIA"))
return GL_VENDOR_NVIDIA;
if (strstr(gl_vendor_string, "ATI"))
return GL_VENDOR_FGLRX;
if (strstr(gl_vendor_string, "Mesa")
|| strstr(gl_vendor_string, "X.Org")
|| strstr(gl_vendor_string, "Advanced Micro Devices, Inc.")
|| strstr(gl_vendor_string, "DRI R300 Project")
|| strstr(gl_vendor_string, "Tungsten Graphics, Inc")
|| strstr(gl_vendor_string, "VMware, Inc.")
|| strstr(gl_vendor_string, "Red Hat")
|| strstr(gl_vendor_string, "Intel")
|| strstr(gl_renderer, "Mesa")
|| strstr(gl_renderer, "Gallium")
|| strstr(gl_renderer, "Intel")
|| strstr(gl_version, "Mesa"))
return GL_VENDOR_MESA;
FIXME("Received unrecognized GL_VENDOR %s. Returning GL_VENDOR_UNKNOWN.\n",
debugstr_a(gl_vendor_string));
return GL_VENDOR_UNKNOWN;
}
static enum wined3d_pci_vendor wined3d_guess_card_vendor(const char *gl_vendor_string, const char *gl_renderer)
{
if (strstr(gl_vendor_string, "NVIDIA")
|| strstr(gl_vendor_string, "Nouveau")
|| strstr(gl_vendor_string, "nouveau"))
return HW_VENDOR_NVIDIA;
if (strstr(gl_vendor_string, "ATI")
|| strstr(gl_vendor_string, "Advanced Micro Devices, Inc.")
|| strstr(gl_vendor_string, "X.Org R300 Project")
|| strstr(gl_renderer, "AMD")
|| strstr(gl_renderer, "FirePro")
|| strstr(gl_renderer, "Radeon")
|| strstr(gl_renderer, "R100")
|| strstr(gl_renderer, "R200")
|| strstr(gl_renderer, "R300")
|| strstr(gl_renderer, "R600")
|| strstr(gl_renderer, "R700"))
return HW_VENDOR_AMD;
if (strstr(gl_vendor_string, "Intel(R)")
/* Intel switched from Intel(R) to Intel® recently, so just match Intel. */
|| strstr(gl_renderer, "Intel")
|| strstr(gl_renderer, "i915")
|| strstr(gl_vendor_string, "Intel Inc."))
return HW_VENDOR_INTEL;
if (strstr(gl_vendor_string, "Red Hat")
|| strstr(gl_renderer, "virgl"))
return HW_VENDOR_REDHAT;
if (strstr(gl_renderer, "SVGA3D"))
return HW_VENDOR_VMWARE;
if (strstr(gl_vendor_string, "Mesa")
|| strstr(gl_vendor_string, "Brian Paul")
|| strstr(gl_vendor_string, "Tungsten Graphics, Inc")
|| strstr(gl_vendor_string, "VMware, Inc."))
return HW_VENDOR_SOFTWARE;
FIXME("Received unrecognized GL_VENDOR %s. Returning HW_VENDOR_NVIDIA.\n", debugstr_a(gl_vendor_string));
return HW_VENDOR_NVIDIA;
}
static enum wined3d_feature_level feature_level_from_caps(const struct wined3d_gl_info *gl_info,
const struct shader_caps *shader_caps, const struct fragment_caps *fragment_caps)
{
unsigned int shader_model;
shader_model = min(shader_caps->vs_version, shader_caps->ps_version);
shader_model = min(shader_model, max(shader_caps->gs_version, 3));
shader_model = min(shader_model, max(shader_caps->hs_version, 4));
shader_model = min(shader_model, max(shader_caps->ds_version, 4));
if (gl_info->supported[WINED3D_GL_VERSION_3_2]
&& gl_info->supported[ARB_POLYGON_OFFSET_CLAMP]
&& gl_info->supported[ARB_SAMPLER_OBJECTS])
{
if (shader_model >= 5
&& gl_info->supported[ARB_DRAW_INDIRECT]
&& gl_info->supported[ARB_TEXTURE_COMPRESSION_BPTC])
return WINED3D_FEATURE_LEVEL_11_1;
if (shader_model >= 4)
{
if (gl_info->supported[ARB_TEXTURE_CUBE_MAP_ARRAY])
return WINED3D_FEATURE_LEVEL_10_1;
return WINED3D_FEATURE_LEVEL_10;
}
}
if (shader_model >= 3 && gl_info->limits.texture_size >= 4096 && gl_info->limits.buffers >= 4)
return WINED3D_FEATURE_LEVEL_9_3;
if (shader_model >= 2)
{
if (gl_info->supported[ARB_OCCLUSION_QUERY]
&& gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE]
&& gl_info->supported[EXT_BLEND_EQUATION_SEPARATE]
&& gl_info->supported[EXT_BLEND_FUNC_SEPARATE])
return WINED3D_FEATURE_LEVEL_9_2;
return WINED3D_FEATURE_LEVEL_9_1;
}
if (shader_model >= 1)
return WINED3D_FEATURE_LEVEL_8;
if (fragment_caps->TextureOpCaps & WINED3DTEXOPCAPS_DOTPRODUCT3)
return WINED3D_FEATURE_LEVEL_7;
if (fragment_caps->MaxSimultaneousTextures > 1)
return WINED3D_FEATURE_LEVEL_6;
return WINED3D_FEATURE_LEVEL_5;
}
static const struct wined3d_renderer_table
{
const char *renderer;
enum wined3d_pci_device id;
}
cards_nvidia_binary[] =
{
/* Direct 3D 11 */
{"RTX 2080 Ti", CARD_NVIDIA_GEFORCE_RTX2080TI}, /* GeForce 2000 - highend */
{"RTX 2080", CARD_NVIDIA_GEFORCE_RTX2080}, /* GeForce 2000 - highend */
{"RTX 2070", CARD_NVIDIA_GEFORCE_RTX2070}, /* GeForce 2000 - highend */
{"RTX 2060", CARD_NVIDIA_GEFORCE_RTX2060}, /* GeForce 2000 - highend */
{"GTX 1660 Ti", CARD_NVIDIA_GEFORCE_GTX1660TI}, /* GeForce 1600 - highend */
{"TITAN V", CARD_NVIDIA_TITANV}, /* GeForce 1000 - highend */
{"TITAN X (Pascal)", CARD_NVIDIA_TITANX_PASCAL}, /* GeForce 1000 - highend */
{"GTX 1080 Ti", CARD_NVIDIA_GEFORCE_GTX1080TI}, /* GeForce 1000 - highend */
{"GTX 1080", CARD_NVIDIA_GEFORCE_GTX1080}, /* GeForce 1000 - highend */
{"GTX 1070", CARD_NVIDIA_GEFORCE_GTX1070}, /* GeForce 1000 - highend */
{"GTX 1060", CARD_NVIDIA_GEFORCE_GTX1060}, /* GeForce 1000 - midend high */
{"GTX 1050 Ti", CARD_NVIDIA_GEFORCE_GTX1050TI}, /* GeForce 1000 - midend */
{"GTX 1050", CARD_NVIDIA_GEFORCE_GTX1050}, /* GeForce 1000 - midend */
{"GTX 980 Ti", CARD_NVIDIA_GEFORCE_GTX980TI}, /* GeForce 900 - highend */
{"GTX 980", CARD_NVIDIA_GEFORCE_GTX980}, /* GeForce 900 - highend */
{"GTX 970M", CARD_NVIDIA_GEFORCE_GTX970M}, /* GeForce 900 - highend mobile*/
{"GTX 970", CARD_NVIDIA_GEFORCE_GTX970}, /* GeForce 900 - highend */
{"GTX TITAN X", CARD_NVIDIA_GEFORCE_GTXTITANX}, /* Geforce 900 - highend */
{"GTX 960M", CARD_NVIDIA_GEFORCE_GTX960M}, /* GeForce 900 - midend high mobile */
{"GTX 960", CARD_NVIDIA_GEFORCE_GTX960}, /* GeForce 900 - midend high */
{"GTX 950M", CARD_NVIDIA_GEFORCE_GTX950M}, /* GeForce 900 - midend mobile */
{"GTX 950", CARD_NVIDIA_GEFORCE_GTX950}, /* GeForce 900 - midend */
{"GeForce 940M", CARD_NVIDIA_GEFORCE_940M}, /* GeForce 900 - midend mobile */
{"GTX 880M", CARD_NVIDIA_GEFORCE_GTX880M}, /* GeForce 800 - mobile */
{"GTX 870M", CARD_NVIDIA_GEFORCE_GTX870M}, /* GeForce 800 - mobile */
{"GTX 860M", CARD_NVIDIA_GEFORCE_GTX860M}, /* GeForce 800 - mobile */
{"GTX 850M", CARD_NVIDIA_GEFORCE_GTX850M}, /* GeForce 800 - mobile */
{"GeForce 845M", CARD_NVIDIA_GEFORCE_845M}, /* GeForce 800 - mobile */
{"GeForce 840M", CARD_NVIDIA_GEFORCE_840M}, /* GeForce 800 - mobile */
{"GeForce 830M", CARD_NVIDIA_GEFORCE_830M}, /* GeForce 800 - mobile */
{"GeForce 820M", CARD_NVIDIA_GEFORCE_820M}, /* GeForce 800 - mobile */
{"GTX 780 Ti", CARD_NVIDIA_GEFORCE_GTX780TI}, /* Geforce 700 - highend */
{"GTX TITAN Black", CARD_NVIDIA_GEFORCE_GTXTITANB}, /* Geforce 700 - highend */
{"GTX TITAN Z", CARD_NVIDIA_GEFORCE_GTXTITANZ}, /* Geforce 700 - highend */
{"GTX TITAN", CARD_NVIDIA_GEFORCE_GTXTITAN}, /* Geforce 700 - highend */
{"GTX 780", CARD_NVIDIA_GEFORCE_GTX780}, /* Geforce 700 - highend */
{"GTX 770M", CARD_NVIDIA_GEFORCE_GTX770M}, /* Geforce 700 - midend high mobile */
{"GTX 770", CARD_NVIDIA_GEFORCE_GTX770}, /* Geforce 700 - highend */
{"GTX 765M", CARD_NVIDIA_GEFORCE_GTX765M}, /* Geforce 700 - midend high mobile */
{"GTX 760 Ti", CARD_NVIDIA_GEFORCE_GTX760TI}, /* Geforce 700 - midend high */
{"GTX 760", CARD_NVIDIA_GEFORCE_GTX760}, /* Geforce 700 - midend high */
{"GTX 750 Ti", CARD_NVIDIA_GEFORCE_GTX750TI}, /* Geforce 700 - midend */
{"GTX 750", CARD_NVIDIA_GEFORCE_GTX750}, /* Geforce 700 - midend */
{"GT 750M", CARD_NVIDIA_GEFORCE_GT750M}, /* Geforce 700 - midend mobile */
{"GT 740M", CARD_NVIDIA_GEFORCE_GT740M}, /* Geforce 700 - midend mobile */
{"GT 730M", CARD_NVIDIA_GEFORCE_GT730M}, /* Geforce 700 - midend mobile */
{"GT 730", CARD_NVIDIA_GEFORCE_GT730}, /* Geforce 700 - lowend */
{"GTX 690", CARD_NVIDIA_GEFORCE_GTX690}, /* Geforce 600 - highend */
{"GTX 680", CARD_NVIDIA_GEFORCE_GTX680}, /* Geforce 600 - highend */
{"GTX 675MX", CARD_NVIDIA_GEFORCE_GTX675MX_1},/* Geforce 600 - highend */
{"GTX 670MX", CARD_NVIDIA_GEFORCE_GTX670MX}, /* Geforce 600 - highend */
{"GTX 670", CARD_NVIDIA_GEFORCE_GTX670}, /* Geforce 600 - midend high */
{"GTX 660 Ti", CARD_NVIDIA_GEFORCE_GTX660TI}, /* Geforce 600 - midend high */
{"GTX 660M", CARD_NVIDIA_GEFORCE_GTX660M}, /* Geforce 600 - midend high mobile */
{"GTX 660", CARD_NVIDIA_GEFORCE_GTX660}, /* Geforce 600 - midend high */
{"GTX 650 Ti", CARD_NVIDIA_GEFORCE_GTX650TI}, /* Geforce 600 - lowend */
{"GTX 650", CARD_NVIDIA_GEFORCE_GTX650}, /* Geforce 600 - lowend */
{"GT 650M", CARD_NVIDIA_GEFORCE_GT650M}, /* Geforce 600 - midend mobile */
{"GT 640M", CARD_NVIDIA_GEFORCE_GT640M}, /* Geforce 600 - midend mobile */
{"GT 630M", CARD_NVIDIA_GEFORCE_GT630M}, /* Geforce 600 - midend mobile */
{"GT 630", CARD_NVIDIA_GEFORCE_GT630}, /* Geforce 600 - lowend */
{"GT 610", CARD_NVIDIA_GEFORCE_GT610}, /* Geforce 600 - lowend */
{"GTX 580", CARD_NVIDIA_GEFORCE_GTX580}, /* Geforce 500 - highend */
{"GTX 570", CARD_NVIDIA_GEFORCE_GTX570}, /* Geforce 500 - midend high */
{"GTX 560 Ti", CARD_NVIDIA_GEFORCE_GTX560TI}, /* Geforce 500 - midend */
{"GTX 560M", CARD_NVIDIA_GEFORCE_GTX560M}, /* Geforce 500 - midend mobile */
{"GTX 560", CARD_NVIDIA_GEFORCE_GTX560}, /* Geforce 500 - midend */
{"GT 555M", CARD_NVIDIA_GEFORCE_GT555M}, /* Geforce 500 - midend mobile */
{"GTX 550 Ti", CARD_NVIDIA_GEFORCE_GTX550}, /* Geforce 500 - midend */
{"GT 540M", CARD_NVIDIA_GEFORCE_GT540M}, /* Geforce 500 - midend mobile */
{"GT 525M", CARD_NVIDIA_GEFORCE_GT525M}, /* Geforce 500 - lowend mobile */
{"GT 520", CARD_NVIDIA_GEFORCE_GT520}, /* Geforce 500 - lowend */
{"GTX 480", CARD_NVIDIA_GEFORCE_GTX480}, /* Geforce 400 - highend */
{"GTX 470", CARD_NVIDIA_GEFORCE_GTX470}, /* Geforce 400 - midend high */
/* Direct 3D 10 */
{"GTX 465", CARD_NVIDIA_GEFORCE_GTX465}, /* Geforce 400 - midend */
{"GTX 460M", CARD_NVIDIA_GEFORCE_GTX460M}, /* Geforce 400 - highend mobile */
{"GTX 460", CARD_NVIDIA_GEFORCE_GTX460}, /* Geforce 400 - midend */
{"GTS 450", CARD_NVIDIA_GEFORCE_GTS450}, /* Geforce 400 - midend low */
{"GT 440", CARD_NVIDIA_GEFORCE_GT440}, /* Geforce 400 - lowend */
{"GT 430", CARD_NVIDIA_GEFORCE_GT430}, /* Geforce 400 - lowend */
{"GT 425M", CARD_NVIDIA_GEFORCE_GT425M}, /* Geforce 400 - lowend mobile */
{"GT 420", CARD_NVIDIA_GEFORCE_GT420}, /* Geforce 400 - lowend */
{"410M", CARD_NVIDIA_GEFORCE_410M}, /* Geforce 400 - lowend mobile */
{"GT 330", CARD_NVIDIA_GEFORCE_GT330}, /* Geforce 300 - highend */
{"GTS 360M", CARD_NVIDIA_GEFORCE_GTS350M}, /* Geforce 300 - highend mobile */
{"GTS 350M", CARD_NVIDIA_GEFORCE_GTS350M}, /* Geforce 300 - highend mobile */
{"GT 330M", CARD_NVIDIA_GEFORCE_GT325M}, /* Geforce 300 - midend mobile */
{"GT 325M", CARD_NVIDIA_GEFORCE_GT325M}, /* Geforce 300 - midend mobile */
{"GT 320M", CARD_NVIDIA_GEFORCE_GT320M}, /* Geforce 300 - midend mobile */
{"320M", CARD_NVIDIA_GEFORCE_320M}, /* Geforce 300 - midend mobile */
{"315M", CARD_NVIDIA_GEFORCE_315M}, /* Geforce 300 - midend mobile */
{"GTX 295", CARD_NVIDIA_GEFORCE_GTX280}, /* Geforce 200 - highend */
{"GTX 285", CARD_NVIDIA_GEFORCE_GTX280}, /* Geforce 200 - highend */
{"GTX 280", CARD_NVIDIA_GEFORCE_GTX280}, /* Geforce 200 - highend */
{"GTX 275", CARD_NVIDIA_GEFORCE_GTX275}, /* Geforce 200 - midend high */
{"GTX 260", CARD_NVIDIA_GEFORCE_GTX260}, /* Geforce 200 - midend */
{"GTS 250", CARD_NVIDIA_GEFORCE_GTS250}, /* Geforce 200 - midend */
{"GT 240", CARD_NVIDIA_GEFORCE_GT240}, /* Geforce 200 - midend */
{"GT 220", CARD_NVIDIA_GEFORCE_GT220}, /* Geforce 200 - lowend */
{"GeForce 310", CARD_NVIDIA_GEFORCE_210}, /* Geforce 200 - lowend */
{"GeForce 305", CARD_NVIDIA_GEFORCE_210}, /* Geforce 200 - lowend */
{"GeForce 210", CARD_NVIDIA_GEFORCE_210}, /* Geforce 200 - lowend */
{"G 210", CARD_NVIDIA_GEFORCE_210}, /* Geforce 200 - lowend */
{"GTS 150", CARD_NVIDIA_GEFORCE_9800GT}, /* Geforce 9 - highend / Geforce 200 - midend */
{"9800", CARD_NVIDIA_GEFORCE_9800GT}, /* Geforce 9 - highend / Geforce 200 - midend */
{"9700M GT", CARD_NVIDIA_GEFORCE_9700MGT}, /* Geforce 9 - midend */
{"GT 140", CARD_NVIDIA_GEFORCE_9600GT}, /* Geforce 9 - midend */
{"9600", CARD_NVIDIA_GEFORCE_9600GT}, /* Geforce 9 - midend */
{"GT 130", CARD_NVIDIA_GEFORCE_9500GT}, /* Geforce 9 - midend low / Geforce 200 - low */
{"GT 120", CARD_NVIDIA_GEFORCE_9500GT}, /* Geforce 9 - midend low / Geforce 200 - low */
{"9500", CARD_NVIDIA_GEFORCE_9500GT}, /* Geforce 9 - midend low / Geforce 200 - low */
{"9400M", CARD_NVIDIA_GEFORCE_9400M}, /* Geforce 9 - lowend */
{"9400", CARD_NVIDIA_GEFORCE_9400GT}, /* Geforce 9 - lowend */
{"9300", CARD_NVIDIA_GEFORCE_9300}, /* Geforce 9 - lowend low */
{"9200", CARD_NVIDIA_GEFORCE_9200}, /* Geforce 9 - lowend low */
{"9100", CARD_NVIDIA_GEFORCE_9200}, /* Geforce 9 - lowend low */
{"G 100", CARD_NVIDIA_GEFORCE_9200}, /* Geforce 9 - lowend low */
{"8800 GTX", CARD_NVIDIA_GEFORCE_8800GTX}, /* Geforce 8 - highend high */
{"8800", CARD_NVIDIA_GEFORCE_8800GTS}, /* Geforce 8 - highend */
{"8600M", CARD_NVIDIA_GEFORCE_8600MGT}, /* Geforce 8 - midend mobile */
{"8600 M", CARD_NVIDIA_GEFORCE_8600MGT}, /* Geforce 8 - midend mobile */
{"8700", CARD_NVIDIA_GEFORCE_8600GT}, /* Geforce 8 - midend */
{"8600", CARD_NVIDIA_GEFORCE_8600GT}, /* Geforce 8 - midend */
{"8500", CARD_NVIDIA_GEFORCE_8500GT}, /* Geforce 8 - mid-lowend */
{"8400", CARD_NVIDIA_GEFORCE_8400GS}, /* Geforce 8 - mid-lowend */
{"8300", CARD_NVIDIA_GEFORCE_8300GS}, /* Geforce 8 - lowend */
{"8200", CARD_NVIDIA_GEFORCE_8200}, /* Geforce 8 - lowend */
{"8100", CARD_NVIDIA_GEFORCE_8200}, /* Geforce 8 - lowend */
/* Direct 3D 9 SM3 */
{"Quadro FX 5", CARD_NVIDIA_GEFORCE_7800GT}, /* Geforce 7 - highend */
{"Quadro FX 4", CARD_NVIDIA_GEFORCE_7800GT}, /* Geforce 7 - highend */
{"7950", CARD_NVIDIA_GEFORCE_7800GT}, /* Geforce 7 - highend */
{"7900", CARD_NVIDIA_GEFORCE_7800GT}, /* Geforce 7 - highend */
{"7800", CARD_NVIDIA_GEFORCE_7800GT}, /* Geforce 7 - highend */
{"7700", CARD_NVIDIA_GEFORCE_7600}, /* Geforce 7 - midend */
{"7600", CARD_NVIDIA_GEFORCE_7600}, /* Geforce 7 - midend */
{"7400", CARD_NVIDIA_GEFORCE_7400}, /* Geforce 7 - lower medium */
{"7300", CARD_NVIDIA_GEFORCE_7300}, /* Geforce 7 - lowend */
{"6800", CARD_NVIDIA_GEFORCE_6800}, /* Geforce 6 - highend */
{"6700", CARD_NVIDIA_GEFORCE_6600GT}, /* Geforce 6 - midend */
{"6610", CARD_NVIDIA_GEFORCE_6600GT}, /* Geforce 6 - midend */
{"6600", CARD_NVIDIA_GEFORCE_6600GT}, /* Geforce 6 - midend */
/* Direct 3D 9 SM2 */
{"Quadro FX", CARD_NVIDIA_GEFORCEFX_5800}, /* GeforceFX - highend */
{"5950", CARD_NVIDIA_GEFORCEFX_5800}, /* GeforceFX - highend */
{"5900", CARD_NVIDIA_GEFORCEFX_5800}, /* GeforceFX - highend */
{"5800", CARD_NVIDIA_GEFORCEFX_5800}, /* GeforceFX - highend */
{"5750", CARD_NVIDIA_GEFORCEFX_5600}, /* GeforceFX - midend */
{"5700", CARD_NVIDIA_GEFORCEFX_5600}, /* GeforceFX - midend */
{"5650", CARD_NVIDIA_GEFORCEFX_5600}, /* GeforceFX - midend */
{"5600", CARD_NVIDIA_GEFORCEFX_5600}, /* GeforceFX - midend */
{"5500", CARD_NVIDIA_GEFORCEFX_5200}, /* GeforceFX - lowend */
{"5300", CARD_NVIDIA_GEFORCEFX_5200}, /* GeforceFX - lowend */
{"5250", CARD_NVIDIA_GEFORCEFX_5200}, /* GeforceFX - lowend */
{"5200", CARD_NVIDIA_GEFORCEFX_5200}, /* GeforceFX - lowend */
{"5100", CARD_NVIDIA_GEFORCEFX_5200}, /* GeforceFX - lowend */
/* Direct 3D 8 */
{"Quadro4", CARD_NVIDIA_GEFORCE4_TI4200},
{"GeForce4 Ti", CARD_NVIDIA_GEFORCE4_TI4200}, /* Geforce4 Ti4200/Ti4400/Ti4600/Ti4800 */
/* Direct 3D 7 */
{"GeForce4 MX", CARD_NVIDIA_GEFORCE4_MX}, /* MX420/MX440/MX460/MX4000 */
{"Quadro2 MXR", CARD_NVIDIA_GEFORCE2_MX},
{"GeForce2 MX", CARD_NVIDIA_GEFORCE2_MX}, /* Geforce2 standard/MX100/MX200/MX400 */
{"Quadro2", CARD_NVIDIA_GEFORCE2},
{"GeForce2", CARD_NVIDIA_GEFORCE2}, /* Geforce2 GTS/Pro/Ti/Ultra */
/* Direct 3D 6 */
{"TNT2", CARD_NVIDIA_RIVA_TNT2}, /* Riva TNT2 standard/M64/Pro/Ultra */
},
/* See http://developer.amd.com/resources/hardware-drivers/ati-catalyst-pc-vendor-id-1002-li/
*
* Beware: renderer string do not match exact card model,
* e.g. HD 4800 is returned for multiple cards, even for RV790 based ones. */
cards_amd_binary[] =
{
{"RX 480", CARD_AMD_RADEON_RX_480},
{"RX 460", CARD_AMD_RADEON_RX_460},
{"R9 Fury Series", CARD_AMD_RADEON_R9_FURY},
/* Southern Islands */
{"HD 7900", CARD_AMD_RADEON_HD7900},
{"HD 7800", CARD_AMD_RADEON_HD7800},
{"HD 7700", CARD_AMD_RADEON_HD7700},
/* Northern Islands */
{"HD 6970", CARD_AMD_RADEON_HD6900},
{"HD 6900", CARD_AMD_RADEON_HD6900},
{"HD 6800", CARD_AMD_RADEON_HD6800},
{"HD 6770M", CARD_AMD_RADEON_HD6600M},
{"HD 6750M", CARD_AMD_RADEON_HD6600M},
{"HD 6700", CARD_AMD_RADEON_HD6700},
{"HD 6670", CARD_AMD_RADEON_HD6600},
{"HD 6630M", CARD_AMD_RADEON_HD6600M},
{"HD 6600M", CARD_AMD_RADEON_HD6600M},
{"HD 6600", CARD_AMD_RADEON_HD6600},
{"HD 6570", CARD_AMD_RADEON_HD6600},
{"HD 6500M", CARD_AMD_RADEON_HD6600M},
{"HD 6500", CARD_AMD_RADEON_HD6600},
{"HD 6480G", CARD_AMD_RADEON_HD6480G},
{"HD 6400", CARD_AMD_RADEON_HD6400},
{"HD 6300", CARD_AMD_RADEON_HD6300},
{"HD 6200", CARD_AMD_RADEON_HD6300},
/* Evergreen */
{"HD 5870", CARD_AMD_RADEON_HD5800}, /* Radeon EG CYPRESS PRO */
{"HD 5850", CARD_AMD_RADEON_HD5800}, /* Radeon EG CYPRESS XT */
{"HD 5800", CARD_AMD_RADEON_HD5800}, /* Radeon EG CYPRESS HD58xx generic renderer string */
{"HD 5770", CARD_AMD_RADEON_HD5700}, /* Radeon EG JUNIPER XT */
{"HD 5750", CARD_AMD_RADEON_HD5700}, /* Radeon EG JUNIPER LE */
{"HD 5700", CARD_AMD_RADEON_HD5700}, /* Radeon EG JUNIPER HD57xx generic renderer string */
{"HD 5670", CARD_AMD_RADEON_HD5600}, /* Radeon EG REDWOOD XT */
{"HD 5570", CARD_AMD_RADEON_HD5600}, /* Radeon EG REDWOOD PRO mapped to HD5600 series */
{"HD 5550", CARD_AMD_RADEON_HD5600}, /* Radeon EG REDWOOD LE mapped to HD5600 series */
{"HD 5450", CARD_AMD_RADEON_HD5400}, /* Radeon EG CEDAR PRO */
{"HD 5000", CARD_AMD_RADEON_HD5600}, /* Defaulting to HD 5600 */
/* R700 */
{"HD 4890", CARD_AMD_RADEON_HD4800}, /* Radeon RV790 */
{"HD 4870", CARD_AMD_RADEON_HD4800}, /* Radeon RV770 */
{"HD 4850", CARD_AMD_RADEON_HD4800}, /* Radeon RV770 */
{"HD 4830", CARD_AMD_RADEON_HD4800}, /* Radeon RV770 */
{"HD 4800", CARD_AMD_RADEON_HD4800}, /* Radeon RV7xx HD48xx generic renderer string */
{"HD 4770", CARD_AMD_RADEON_HD4700}, /* Radeon RV740 */
{"HD 4700", CARD_AMD_RADEON_HD4700}, /* Radeon RV7xx HD47xx generic renderer string */
{"HD 4670", CARD_AMD_RADEON_HD4600}, /* Radeon RV730 */
{"HD 4650", CARD_AMD_RADEON_HD4600}, /* Radeon RV730 */
{"HD 4600", CARD_AMD_RADEON_HD4600}, /* Radeon RV730 */
{"HD 4550", CARD_AMD_RADEON_HD4350}, /* Radeon RV710 */
{"HD 4350", CARD_AMD_RADEON_HD4350}, /* Radeon RV710 */
/* R600/R700 integrated */
{"HD 4200M", CARD_AMD_RADEON_HD4200M},
{"HD 3300", CARD_AMD_RADEON_HD3200},
{"HD 3200", CARD_AMD_RADEON_HD3200},
{"HD 3100", CARD_AMD_RADEON_HD3200},
/* R600 */
{"HD 3870", CARD_AMD_RADEON_HD2900}, /* HD2900/HD3800 - highend */
{"HD 3850", CARD_AMD_RADEON_HD2900}, /* HD2900/HD3800 - highend */
{"HD 2900", CARD_AMD_RADEON_HD2900}, /* HD2900/HD3800 - highend */
{"HD 3830", CARD_AMD_RADEON_HD2600}, /* China-only midend */
{"HD 3690", CARD_AMD_RADEON_HD2600}, /* HD2600/HD3600 - midend */
{"HD 3650", CARD_AMD_RADEON_HD2600}, /* HD2600/HD3600 - midend */
{"HD 2600", CARD_AMD_RADEON_HD2600}, /* HD2600/HD3600 - midend */
{"HD 3470", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
{"HD 3450", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
{"HD 3430", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
{"HD 3400", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
{"HD 2400", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
{"HD 2350", CARD_AMD_RADEON_HD2350}, /* HD2350/HD2400/HD3400 - lowend */
/* Radeon R5xx */
{"X1950", CARD_AMD_RADEON_X1600},
{"X1900", CARD_AMD_RADEON_X1600},
{"X1800", CARD_AMD_RADEON_X1600},
{"X1650", CARD_AMD_RADEON_X1600},
{"X1600", CARD_AMD_RADEON_X1600},
/* Radeon R4xx + X1300/X1400/X1450/X1550/X2300/X2500/HD2300 (lowend R5xx)
* Note X2300/X2500/HD2300 are R5xx GPUs with a 2xxx naming but they are still DX9-only */
{"HD 2300", CARD_AMD_RADEON_X700},
{"X2500", CARD_AMD_RADEON_X700},
{"X2300", CARD_AMD_RADEON_X700},
{"X1550", CARD_AMD_RADEON_X700},
{"X1450", CARD_AMD_RADEON_X700},
{"X1400", CARD_AMD_RADEON_X700},
{"X1300", CARD_AMD_RADEON_X700},
{"X850", CARD_AMD_RADEON_X700},
{"X800", CARD_AMD_RADEON_X700},
{"X700", CARD_AMD_RADEON_X700},
/* Radeon Xpress Series - onboard, DX9b, Shader 2.0, 300-400 MHz */
{"Radeon Xpress", CARD_AMD_RADEON_XPRESS_200M},
},
cards_intel[] =
{
/* Skylake */
{"Iris Pro Graphics P580", CARD_INTEL_IPP580_1},
{"Skylake", CARD_INTEL_HD520_1},
/* Broadwell */
{"Iris Pro P6300", CARD_INTEL_IPP6300},
{"Iris Pro 6200", CARD_INTEL_IP6200},
{"Iris 6100", CARD_INTEL_I6100},
{"Iris(TM) Graphics 6100", CARD_INTEL_I6100}, /* MacOS */
/* Haswell */
{"Iris Pro 5200", CARD_INTEL_IP5200_1},
{"Iris 5100", CARD_INTEL_I5100_1},
{"HD Graphics 5000", CARD_INTEL_HD5000_1}, /* MacOS */
{"Haswell Mobile", CARD_INTEL_HWM},
{"Iris OpenGL Engine", CARD_INTEL_HWM}, /* MacOS */
/* Ivybridge */
{"Ivybridge Server", CARD_INTEL_IVBS},
{"Ivybridge Mobile", CARD_INTEL_IVBM},
{"Ivybridge Desktop", CARD_INTEL_IVBD},
{"HD Graphics 4000", CARD_INTEL_IVBD}, /* MacOS */
/* Sandybridge */
{"Sandybridge Server", CARD_INTEL_SNBS},
{"Sandybridge Mobile", CARD_INTEL_SNBM},
{"Sandybridge Desktop", CARD_INTEL_SNBD},
/* Ironlake */
{"Ironlake Mobile", CARD_INTEL_ILKM},
{"Ironlake Desktop", CARD_INTEL_ILKD},
/* G4x */
{"B43", CARD_INTEL_B43},
{"G41", CARD_INTEL_G41},
{"G45", CARD_INTEL_G45},
{"Q45", CARD_INTEL_Q45},
{"Integrated Graphics Device", CARD_INTEL_IGD},
{"GM45", CARD_INTEL_GM45},
/* i965 */
{"965GME", CARD_INTEL_965GME},
{"965GM", CARD_INTEL_965GM},
{"X3100", CARD_INTEL_965GM}, /* MacOS */
{"946GZ", CARD_INTEL_946GZ},
{"965G", CARD_INTEL_965G},
{"965Q", CARD_INTEL_965Q},
/* i945 */
{"Pineview M", CARD_INTEL_PNVM},
{"Pineview G", CARD_INTEL_PNVG},
{"IGD", CARD_INTEL_PNVG},
{"Q33", CARD_INTEL_Q33},
{"G33", CARD_INTEL_G33},
{"Q35", CARD_INTEL_Q35},
{"945GME", CARD_INTEL_945GME},
{"945GM", CARD_INTEL_945GM},
{"GMA 950", CARD_INTEL_945GM}, /* MacOS */
{"945G", CARD_INTEL_945G},
/* i915 */
{"915GM", CARD_INTEL_915GM},
{"E7221G", CARD_INTEL_E7221G},
{"915G", CARD_INTEL_915G},
/* i8xx */
{"865G", CARD_INTEL_865G},
{"845G", CARD_INTEL_845G},
{"855GM", CARD_INTEL_855GM},
{"830M", CARD_INTEL_830M},
},
/* 20101109 - These are never returned by current Gallium radeon
* drivers: R700, RV790, R680, RV535, RV516, R410, RS485, RV360, RV351. */
cards_amd_mesa[] =
{
/* Polaris 10/11 */
{"POLARIS10", CARD_AMD_RADEON_RX_480},
{"POLARIS11", CARD_AMD_RADEON_RX_460},
/* Volcanic Islands */
{"FIJI", CARD_AMD_RADEON_R9_FURY},
{"TONGA", CARD_AMD_RADEON_R9_285},
/* Sea Islands */
{"HAWAII", CARD_AMD_RADEON_R9_290},
{"KAVERI", CARD_AMD_RADEON_R7 },
{"KABINI", CARD_AMD_RADEON_R3 },
{"BONAIRE", CARD_AMD_RADEON_HD8770},
/* Southern Islands */
{"OLAND", CARD_AMD_RADEON_HD8670},
{"HAINAN", CARD_AMD_RADEON_HD8600M},
{"TAHITI", CARD_AMD_RADEON_HD7900},
{"PITCAIRN", CARD_AMD_RADEON_HD7800},
{"CAPE VERDE", CARD_AMD_RADEON_HD7700},
/* Northern Islands */
{"ARUBA", CARD_AMD_RADEON_HD7660D},
{"CAYMAN", CARD_AMD_RADEON_HD6900},
{"BARTS", CARD_AMD_RADEON_HD6800},
{"TURKS", CARD_AMD_RADEON_HD6600},
{"SUMO2", CARD_AMD_RADEON_HD6410D}, /* SUMO2 first, because we do a strstr(). */
{"SUMO", CARD_AMD_RADEON_HD6550D},
{"CAICOS", CARD_AMD_RADEON_HD6400},
{"PALM", CARD_AMD_RADEON_HD6300},
/* Evergreen */
{"HEMLOCK", CARD_AMD_RADEON_HD5900},
{"CYPRESS", CARD_AMD_RADEON_HD5800},
{"JUNIPER", CARD_AMD_RADEON_HD5700},
{"REDWOOD", CARD_AMD_RADEON_HD5600},
{"CEDAR", CARD_AMD_RADEON_HD5400},
/* R700 */
{"R700", CARD_AMD_RADEON_HD4800},
{"RV790", CARD_AMD_RADEON_HD4800},
{"RV770", CARD_AMD_RADEON_HD4800},
{"RV740", CARD_AMD_RADEON_HD4700},
{"RV730", CARD_AMD_RADEON_HD4600},
{"RV710", CARD_AMD_RADEON_HD4350},
/* R600/R700 integrated */
{"RS880", CARD_AMD_RADEON_HD4200M},
{"RS780", CARD_AMD_RADEON_HD3200},
/* R600 */
{"R680", CARD_AMD_RADEON_HD2900},
{"R600", CARD_AMD_RADEON_HD2900},
{"RV670", CARD_AMD_RADEON_HD3850},
{"RV635", CARD_AMD_RADEON_HD2600},
{"RV630", CARD_AMD_RADEON_HD2600},
{"RV620", CARD_AMD_RADEON_HD2350},
{"RV610", CARD_AMD_RADEON_HD2350},
/* R500 */
{"R580", CARD_AMD_RADEON_X1600},
{"R520", CARD_AMD_RADEON_X1600},
{"RV570", CARD_AMD_RADEON_X1600},
{"RV560", CARD_AMD_RADEON_X1600},
{"RV535", CARD_AMD_RADEON_X1600},
{"RV530", CARD_AMD_RADEON_X1600},
{"RV516", CARD_AMD_RADEON_X700},
{"RV515", CARD_AMD_RADEON_X700},
/* R400 */
{"R481", CARD_AMD_RADEON_X700},
{"R480", CARD_AMD_RADEON_X700},
{"R430", CARD_AMD_RADEON_X700},
{"R423", CARD_AMD_RADEON_X700},
{"R420", CARD_AMD_RADEON_X700},
{"R410", CARD_AMD_RADEON_X700},
{"RV410", CARD_AMD_RADEON_X700},
/* Radeon Xpress - onboard, DX9b, Shader 2.0, 300-400 MHz */
{"RS740", CARD_AMD_RADEON_XPRESS_200M},
{"RS690", CARD_AMD_RADEON_XPRESS_200M},
{"RS600", CARD_AMD_RADEON_XPRESS_200M},
{"RS485", CARD_AMD_RADEON_XPRESS_200M},
{"RS482", CARD_AMD_RADEON_XPRESS_200M},
{"RS480", CARD_AMD_RADEON_XPRESS_200M},
{"RS400", CARD_AMD_RADEON_XPRESS_200M},
{"RC410", CARD_AMD_RADEON_XPRESS_200M},
/* R300 */
{"R360", CARD_AMD_RADEON_9500},
{"R350", CARD_AMD_RADEON_9500},
{"R300", CARD_AMD_RADEON_9500},
{"RV380", CARD_AMD_RADEON_9500},
{"RV370", CARD_AMD_RADEON_9500},
{"RV360", CARD_AMD_RADEON_9500},
{"RV351", CARD_AMD_RADEON_9500},
{"RV350", CARD_AMD_RADEON_9500},
},
cards_nvidia_mesa[] =
{
/* Maxwell */
{"NV124", CARD_NVIDIA_GEFORCE_GTX970},
{"NV120", CARD_NVIDIA_GEFORCE_GTX980TI},
{"NV118", CARD_NVIDIA_GEFORCE_840M},
{"NV117", CARD_NVIDIA_GEFORCE_GTX750},
/* Kepler */
{"NV108", CARD_NVIDIA_GEFORCE_GT740M},
{"NV106", CARD_NVIDIA_GEFORCE_GT720},
{"NVF1", CARD_NVIDIA_GEFORCE_GTX780TI},
{"NVF0", CARD_NVIDIA_GEFORCE_GTX780},
{"NVE6", CARD_NVIDIA_GEFORCE_GTX770M},
{"NVE4", CARD_NVIDIA_GEFORCE_GTX680}, /* 690 / 675MX / 760TI */
/* Fermi */
{"NVD9", CARD_NVIDIA_GEFORCE_GT520},
{"NVD7", CARD_NVIDIA_GEFORCE_820M},
{"NVCF", CARD_NVIDIA_GEFORCE_GTX550},
{"NVCE", CARD_NVIDIA_GEFORCE_GTX560},
{"NVC8", CARD_NVIDIA_GEFORCE_GTX570},
{"NVC4", CARD_NVIDIA_GEFORCE_GTX460},
{"NVC3", CARD_NVIDIA_GEFORCE_GT440},
{"NVC1", CARD_NVIDIA_GEFORCE_GT420},
{"NVC0", CARD_NVIDIA_GEFORCE_GTX480},
/* Tesla */
{"NVAF", CARD_NVIDIA_GEFORCE_GT320M},
{"NVAC", CARD_NVIDIA_GEFORCE_8200},
{"NVAA", CARD_NVIDIA_GEFORCE_8200}, /* 8100 */
{"NVA8", CARD_NVIDIA_GEFORCE_210},
{"NVA5", CARD_NVIDIA_GEFORCE_GT220},
{"NVA3", CARD_NVIDIA_GEFORCE_GT240},
{"NVA0", CARD_NVIDIA_GEFORCE_GTX280},
{"NV98", CARD_NVIDIA_GEFORCE_9200},
{"NV96", CARD_NVIDIA_GEFORCE_9400GT},
{"NV94", CARD_NVIDIA_GEFORCE_9600GT},
{"NV92", CARD_NVIDIA_GEFORCE_9800GT},
{"NV86", CARD_NVIDIA_GEFORCE_8500GT},
{"NV84", CARD_NVIDIA_GEFORCE_8600GT},
{"NV50", CARD_NVIDIA_GEFORCE_8800GTX},
/* Curie */
{"NV68", CARD_NVIDIA_GEFORCE_6200}, /* 7050 */
{"NV67", CARD_NVIDIA_GEFORCE_6200}, /* 7000M */
{"NV63", CARD_NVIDIA_GEFORCE_6200}, /* 7100 */
{"NV4E", CARD_NVIDIA_GEFORCE_6200}, /* 6100 Go / 6150 Go */
{"NV4C", CARD_NVIDIA_GEFORCE_6200}, /* 6150SE */
{"NV4B", CARD_NVIDIA_GEFORCE_7600},
{"NV4A", CARD_NVIDIA_GEFORCE_6200},
{"NV49", CARD_NVIDIA_GEFORCE_7800GT}, /* 7900 */
{"NV47", CARD_NVIDIA_GEFORCE_7800GT},
{"NV46", CARD_NVIDIA_GEFORCE_7400},
{"NV45", CARD_NVIDIA_GEFORCE_6800},
{"NV44", CARD_NVIDIA_GEFORCE_6200},
{"NV43", CARD_NVIDIA_GEFORCE_6600GT},
{"NV42", CARD_NVIDIA_GEFORCE_6800},
{"NV41", CARD_NVIDIA_GEFORCE_6800},
{"NV40", CARD_NVIDIA_GEFORCE_6800},
/* Rankine */
{"NV38", CARD_NVIDIA_GEFORCEFX_5800}, /* FX 5950 Ultra */
{"NV36", CARD_NVIDIA_GEFORCEFX_5800}, /* FX 5700/5750 */
{"NV35", CARD_NVIDIA_GEFORCEFX_5800}, /* FX 5900 */
{"NV34", CARD_NVIDIA_GEFORCEFX_5200},
{"NV31", CARD_NVIDIA_GEFORCEFX_5600},
{"NV30", CARD_NVIDIA_GEFORCEFX_5800},
/* Kelvin */
{"nv28", CARD_NVIDIA_GEFORCE4_TI4200},
{"nv25", CARD_NVIDIA_GEFORCE4_TI4200},
{"nv20", CARD_NVIDIA_GEFORCE3},
/* Celsius */
{"nv1F", CARD_NVIDIA_GEFORCE4_MX}, /* GF4 MX IGP */
{"nv1A", CARD_NVIDIA_GEFORCE2}, /* GF2 IGP */
{"nv18", CARD_NVIDIA_GEFORCE4_MX},
{"nv17", CARD_NVIDIA_GEFORCE4_MX},
{"nv16", CARD_NVIDIA_GEFORCE2},
{"nv15", CARD_NVIDIA_GEFORCE2},
{"nv11", CARD_NVIDIA_GEFORCE2_MX},
{"nv10", CARD_NVIDIA_GEFORCE},
/* Fahrenheit */
{"nv05", CARD_NVIDIA_RIVA_TNT2},
{"nv04", CARD_NVIDIA_RIVA_TNT},
{"nv03", CARD_NVIDIA_RIVA_128},
},
cards_redhat[] =
{
{"virgl", CARD_REDHAT_VIRGL},
},
cards_vmware[] =
{
{"SVGA3D", CARD_VMWARE_SVGA3D},
};
static const struct gl_vendor_selection
{
enum wined3d_gl_vendor gl_vendor;
const char *description; /* Description of the card selector i.e. Apple OS/X Intel */
const struct wined3d_renderer_table *cards; /* To be used as cards[], pointer to the first member in an array */
size_t cards_size; /* Number of entries in the array above */
}
amd_gl_vendor_table[] =
{
{GL_VENDOR_APPLE, "Apple OSX AMD/ATI binary driver", cards_amd_binary, ARRAY_SIZE(cards_amd_binary)},
{GL_VENDOR_FGLRX, "AMD/ATI binary driver", cards_amd_binary, ARRAY_SIZE(cards_amd_binary)},
{GL_VENDOR_MESA, "Mesa AMD/ATI driver", cards_amd_mesa, ARRAY_SIZE(cards_amd_mesa)},
},
nvidia_gl_vendor_table[] =
{
{GL_VENDOR_APPLE, "Apple OSX NVidia binary driver", cards_nvidia_binary, ARRAY_SIZE(cards_nvidia_binary)},
{GL_VENDOR_MESA, "Mesa Nouveau driver", cards_nvidia_mesa, ARRAY_SIZE(cards_nvidia_mesa)},
{GL_VENDOR_NVIDIA, "NVIDIA binary driver", cards_nvidia_binary, ARRAY_SIZE(cards_nvidia_binary)},
},
redhat_gl_vendor_table[] =
{
{GL_VENDOR_MESA, "Red Hat driver", cards_redhat, ARRAY_SIZE(cards_redhat)},
},
vmware_gl_vendor_table[] =
{
{GL_VENDOR_MESA, "VMware driver", cards_vmware, ARRAY_SIZE(cards_vmware)},
},
intel_gl_vendor_table[] =
{
{GL_VENDOR_APPLE, "Apple OSX Intel binary driver", cards_intel, ARRAY_SIZE(cards_intel)},
{GL_VENDOR_MESA, "Mesa Intel driver", cards_intel, ARRAY_SIZE(cards_intel)},
};
static enum wined3d_pci_device select_card_handler(const struct gl_vendor_selection *table,
unsigned int table_size, enum wined3d_gl_vendor gl_vendor, const char *gl_renderer)
{
unsigned int i, j;
for (i = 0; i < table_size; ++i)
{
if (table[i].gl_vendor != gl_vendor)
continue;
TRACE("Applying card selector \"%s\".\n", table[i].description);
for (j = 0; j < table[i].cards_size; ++j)
{
if (strstr(gl_renderer, table[i].cards[j].renderer))
return table[i].cards[j].id;
}
return PCI_DEVICE_NONE;
}
FIXME("Couldn't find a suitable card selector for GL vendor %04x (using GL_RENDERER %s)\n",
gl_vendor, debugstr_a(gl_renderer));
return PCI_DEVICE_NONE;
}
static const struct
{
enum wined3d_pci_vendor card_vendor;
const char *description; /* Description of the card selector i.e. Apple OS/X Intel */
const struct gl_vendor_selection *gl_vendor_selection;
unsigned int gl_vendor_count;
}
card_vendor_table[] =
{
{HW_VENDOR_AMD, "AMD", amd_gl_vendor_table, ARRAY_SIZE(amd_gl_vendor_table)},
{HW_VENDOR_NVIDIA, "NVIDIA", nvidia_gl_vendor_table, ARRAY_SIZE(nvidia_gl_vendor_table)},
{HW_VENDOR_REDHAT, "Red Hat",redhat_gl_vendor_table, ARRAY_SIZE(redhat_gl_vendor_table)},
{HW_VENDOR_VMWARE, "VMware", vmware_gl_vendor_table, ARRAY_SIZE(vmware_gl_vendor_table)},
{HW_VENDOR_INTEL, "Intel", intel_gl_vendor_table, ARRAY_SIZE(intel_gl_vendor_table)},
};
static enum wined3d_pci_device wined3d_guess_card(enum wined3d_feature_level feature_level,
const char *gl_renderer, enum wined3d_gl_vendor *gl_vendor,
enum wined3d_pci_vendor *card_vendor)
{
/* A Direct3D device object contains the PCI id (vendor + device) of the
* videocard which is used for rendering. Various applications use this
* information to get a rough estimation of the features of the card and
* some might use it for enabling 3d effects only on certain types of
* videocards. In some cases games might even use it to work around bugs
* which happen on certain videocards/driver combinations. The problem is
* that OpenGL only exposes a rendering string containing the name of the
* videocard and not the PCI id.
*
* Various games depend on the PCI id, so somehow we need to provide one.
* A simple option is to parse the renderer string and translate this to
* the right PCI id. This is a lot of work because there are more than 200
* GPUs just for NVIDIA. Various cards share the same renderer string, so
* the amount of code might be 'small' but there are quite a number of
* exceptions which would make this a pain to maintain. Another way would
* be to query the PCI id from the operating system (assuming this is the
* videocard which is used for rendering which is not always the case).
* This would work but it is not very portable. Second it would not work
* well in, let's say, a remote X situation in which the amount of 3d
* features which can be used is limited.
*
* As said most games only use the PCI id to get an indication of the
* capabilities of the card. It doesn't really matter if the given id is
* the correct one if we return the id of a card with similar 3d features.
*
* The code below checks the OpenGL capabilities of a videocard and matches
* that to a certain level of Direct3D functionality. Once a card passes
* the Direct3D9 check, we know that the card (in case of NVIDIA) is at
* least a GeforceFX. To give a better estimate we do a basic check on the
* renderer string but if that won't pass we return a default card. This
* way is better than maintaining a full card database as even without a
* full database we can return a card with similar features. Second the
* size of the database can be made quite small because when you know what
* type of 3d functionality a card has, you know to which GPU family the
* GPU must belong. Because of this you only have to check a small part of
* the renderer string to distinguish between different models from that
* family.
*
* The code also selects a default amount of video memory which we will
* use for an estimation of the amount of free texture memory. In case of
* real D3D the amount of texture memory includes video memory and system
* memory (to be specific AGP memory or in case of PCIE TurboCache /
* HyperMemory). We don't know how much system memory can be addressed by
* the system but we can make a reasonable estimation about the amount of
* video memory. If the value is slightly wrong it doesn't matter as we
* didn't include AGP-like memory which makes the amount of addressable
* memory higher and second OpenGL isn't that critical it moves to system
* memory behind our backs if really needed. Note that the amount of video
* memory can be overruled using a registry setting. */
enum wined3d_pci_device device;
unsigned int i;
for (i = 0; i < ARRAY_SIZE(card_vendor_table); ++i)
{
if (card_vendor_table[i].card_vendor != *card_vendor)
continue;
TRACE("Applying card selector \"%s\".\n", card_vendor_table[i].description);
device = select_card_handler(card_vendor_table[i].gl_vendor_selection,
card_vendor_table[i].gl_vendor_count, *gl_vendor, gl_renderer);
if (device != PCI_DEVICE_NONE)
return device;
TRACE("Unrecognized renderer %s, falling back to default.\n", debugstr_a(gl_renderer));
return wined3d_gpu_from_feature_level(card_vendor, feature_level);
}
FIXME("No card selector available for card vendor %04x (using GL_RENDERER %s).\n",
*card_vendor, debugstr_a(gl_renderer));
return wined3d_gpu_from_feature_level(card_vendor, feature_level);
}
static const struct wined3d_vertex_pipe_ops *select_vertex_implementation(const struct wined3d_gl_info *gl_info,
const struct wined3d_shader_backend_ops *shader_backend_ops)
{
if (shader_backend_ops == &glsl_shader_backend && gl_info->supported[ARB_VERTEX_SHADER])
return &glsl_vertex_pipe;
return &ffp_vertex_pipe;
}
static const struct wined3d_fragment_pipe_ops *select_fragment_implementation(const struct wined3d_gl_info *gl_info,
const struct wined3d_shader_backend_ops *shader_backend_ops)
{
if (shader_backend_ops == &glsl_shader_backend && gl_info->supported[ARB_FRAGMENT_SHADER])
return &glsl_fragment_pipe;
if (gl_info->supported[ARB_FRAGMENT_PROGRAM])
return &arbfp_fragment_pipeline;
if (gl_info->supported[ATI_FRAGMENT_SHADER])
return &atifs_fragment_pipeline;
if (gl_info->supported[NV_REGISTER_COMBINERS] && gl_info->supported[NV_TEXTURE_SHADER2])
return &nvts_fragment_pipeline;
if (gl_info->supported[NV_REGISTER_COMBINERS])
return &nvrc_fragment_pipeline;
return &ffp_fragment_pipeline;
}
static const struct wined3d_shader_backend_ops *select_shader_backend(const struct wined3d_gl_info *gl_info)
{
BOOL glsl = wined3d_settings.shader_backend == WINED3D_SHADER_BACKEND_AUTO
|| wined3d_settings.shader_backend == WINED3D_SHADER_BACKEND_GLSL;
BOOL arb = wined3d_settings.shader_backend == WINED3D_SHADER_BACKEND_AUTO
|| wined3d_settings.shader_backend == WINED3D_SHADER_BACKEND_ARB;
if (!gl_info->supported[WINED3D_GL_LEGACY_CONTEXT] && !glsl)
{
ERR_(winediag)("Ignoring the shader backend registry key. "
"GLSL is the only shader backend available on core profile contexts. "
"You need to explicitly set GL version to use legacy contexts.\n");
glsl = TRUE;
}
glsl = glsl && gl_info->glsl_version >= MAKEDWORD_VERSION(1, 20);
if (glsl && gl_info->supported[ARB_VERTEX_SHADER] && gl_info->supported[ARB_FRAGMENT_SHADER])
return &glsl_shader_backend;
if (arb && gl_info->supported[ARB_VERTEX_PROGRAM] && gl_info->supported[ARB_FRAGMENT_PROGRAM])
return &arb_program_shader_backend;
if (glsl && (gl_info->supported[ARB_VERTEX_SHADER] || gl_info->supported[ARB_FRAGMENT_SHADER]))
return &glsl_shader_backend;
if (arb && (gl_info->supported[ARB_VERTEX_PROGRAM] || gl_info->supported[ARB_FRAGMENT_PROGRAM]))
return &arb_program_shader_backend;
return &none_shader_backend;
}
static void parse_extension_string(struct wined3d_gl_info *gl_info, const char *extensions,
const struct wined3d_extension_map *map, UINT entry_count)
{
while (*extensions)
{
const char *start;
size_t len;
UINT i;
while (isspace(*extensions))
++extensions;
start = extensions;
while (!isspace(*extensions) && *extensions)
++extensions;
len = extensions - start;
if (!len)
continue;
TRACE("- %s.\n", debugstr_an(start, len));
for (i = 0; i < entry_count; ++i)
{
if (len == strlen(map[i].extension_string)
&& !memcmp(start, map[i].extension_string, len))
{
TRACE(" FOUND: %s support.\n", map[i].extension_string);
gl_info->supported[map[i].extension] = TRUE;
break;
}
}
}
}
static void enumerate_gl_extensions(struct wined3d_gl_info *gl_info,
const struct wined3d_extension_map *map, unsigned int map_entries_count)
{
const char *gl_extension_name;
unsigned int i, j;
GLint extensions_count;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_NUM_EXTENSIONS, &extensions_count);
for (i = 0; i < extensions_count; ++i)
{
gl_extension_name = (const char *)GL_EXTCALL(glGetStringi(GL_EXTENSIONS, i));
TRACE("- %s.\n", debugstr_a(gl_extension_name));
for (j = 0; j < map_entries_count; ++j)
{
if (!strcmp(gl_extension_name, map[j].extension_string))
{
TRACE("FOUND: %s support.\n", map[j].extension_string);
gl_info->supported[map[j].extension] = TRUE;
break;
}
}
}
}
static void load_gl_funcs(struct wined3d_gl_info *gl_info)
{
#define USE_GL_FUNC(pfn) gl_info->gl_ops.ext.p_##pfn = (void *)wglGetProcAddress(#pfn);
/* GL_APPLE_fence */
USE_GL_FUNC(glDeleteFencesAPPLE)
USE_GL_FUNC(glFinishFenceAPPLE)
USE_GL_FUNC(glFinishObjectAPPLE)
USE_GL_FUNC(glGenFencesAPPLE)
USE_GL_FUNC(glIsFenceAPPLE)
USE_GL_FUNC(glSetFenceAPPLE)
USE_GL_FUNC(glTestFenceAPPLE)
USE_GL_FUNC(glTestObjectAPPLE)
/* GL_APPLE_flush_buffer_range */
USE_GL_FUNC(glBufferParameteriAPPLE)
USE_GL_FUNC(glFlushMappedBufferRangeAPPLE)
/* GL_ARB_base_instance */
USE_GL_FUNC(glDrawArraysInstancedBaseInstance)
USE_GL_FUNC(glDrawElementsInstancedBaseVertexBaseInstance)
/* GL_ARB_blend_func_extended */
USE_GL_FUNC(glBindFragDataLocationIndexed)
USE_GL_FUNC(glGetFragDataIndex)
/* GL_ARB_buffer_storage */
USE_GL_FUNC(glBufferStorage)
/* GL_ARB_clear_buffer_object */
USE_GL_FUNC(glClearBufferData)
USE_GL_FUNC(glClearBufferSubData)
/* GL_ARB_clear_texture */
USE_GL_FUNC(glClearTexImage)
USE_GL_FUNC(glClearTexSubImage)
/* GL_ARB_clip_control */
USE_GL_FUNC(glClipControl)
/* GL_ARB_color_buffer_float */
USE_GL_FUNC(glClampColorARB)
/* GL_ARB_compute_shader */
USE_GL_FUNC(glDispatchCompute)
USE_GL_FUNC(glDispatchComputeIndirect)
/* GL_ARB_copy_buffer */
USE_GL_FUNC(glCopyBufferSubData)
/* GL_ARB_copy_image */
USE_GL_FUNC(glCopyImageSubData)
/* GL_ARB_debug_output */
USE_GL_FUNC(glDebugMessageCallbackARB)
USE_GL_FUNC(glDebugMessageControlARB)
USE_GL_FUNC(glDebugMessageInsertARB)
USE_GL_FUNC(glGetDebugMessageLogARB)
/* GL_ARB_draw_buffers */
USE_GL_FUNC(glDrawBuffersARB)
/* GL_ARB_draw_elements_base_vertex */
USE_GL_FUNC(glDrawElementsBaseVertex)
USE_GL_FUNC(glDrawElementsInstancedBaseVertex)
USE_GL_FUNC(glDrawRangeElementsBaseVertex)
USE_GL_FUNC(glMultiDrawElementsBaseVertex)
/* GL_ARB_draw_indirect */
USE_GL_FUNC(glDrawArraysIndirect)
USE_GL_FUNC(glDrawElementsIndirect)
/* GL_ARB_draw_instanced */
USE_GL_FUNC(glDrawArraysInstancedARB)
USE_GL_FUNC(glDrawElementsInstancedARB)
/* GL_ARB_ES2_compatibility */
USE_GL_FUNC(glReleaseShaderCompiler)
USE_GL_FUNC(glShaderBinary)
USE_GL_FUNC(glGetShaderPrecisionFormat)
USE_GL_FUNC(glDepthRangef)
USE_GL_FUNC(glClearDepthf)
/* GL_ARB_framebuffer_no_attachments */
USE_GL_FUNC(glFramebufferParameteri)
/* GL_ARB_framebuffer_object */
USE_GL_FUNC(glBindFramebuffer)
USE_GL_FUNC(glBindRenderbuffer)
USE_GL_FUNC(glBlitFramebuffer)
USE_GL_FUNC(glCheckFramebufferStatus)
USE_GL_FUNC(glDeleteFramebuffers)
USE_GL_FUNC(glDeleteRenderbuffers)
USE_GL_FUNC(glFramebufferRenderbuffer)
USE_GL_FUNC(glFramebufferTexture)
USE_GL_FUNC(glFramebufferTexture1D)
USE_GL_FUNC(glFramebufferTexture2D)
USE_GL_FUNC(glFramebufferTexture3D)
USE_GL_FUNC(glFramebufferTextureLayer)
USE_GL_FUNC(glGenFramebuffers)
USE_GL_FUNC(glGenRenderbuffers)
USE_GL_FUNC(glGenerateMipmap)
USE_GL_FUNC(glGetFramebufferAttachmentParameteriv)
USE_GL_FUNC(glGetRenderbufferParameteriv)
USE_GL_FUNC(glIsFramebuffer)
USE_GL_FUNC(glIsRenderbuffer)
USE_GL_FUNC(glRenderbufferStorage)
USE_GL_FUNC(glRenderbufferStorageMultisample)
/* GL_ARB_geometry_shader4 */
USE_GL_FUNC(glFramebufferTextureARB)
USE_GL_FUNC(glFramebufferTextureFaceARB)
USE_GL_FUNC(glFramebufferTextureLayerARB)
USE_GL_FUNC(glProgramParameteriARB)
/* GL_ARB_instanced_arrays */
USE_GL_FUNC(glVertexAttribDivisorARB)
/* GL_ARB_internalformat_query */
USE_GL_FUNC(glGetInternalformativ)
/* GL_ARB_internalformat_query2 */
USE_GL_FUNC(glGetInternalformati64v)
/* GL_ARB_map_buffer_range */
USE_GL_FUNC(glFlushMappedBufferRange)
USE_GL_FUNC(glMapBufferRange)
/* GL_ARB_multisample */
USE_GL_FUNC(glSampleCoverageARB)
/* GL_ARB_multitexture */
USE_GL_FUNC(glActiveTextureARB)
USE_GL_FUNC(glClientActiveTextureARB)
USE_GL_FUNC(glMultiTexCoord1fARB)
USE_GL_FUNC(glMultiTexCoord1fvARB)
USE_GL_FUNC(glMultiTexCoord2fARB)
USE_GL_FUNC(glMultiTexCoord2fvARB)
USE_GL_FUNC(glMultiTexCoord2svARB)
USE_GL_FUNC(glMultiTexCoord3fARB)
USE_GL_FUNC(glMultiTexCoord3fvARB)
USE_GL_FUNC(glMultiTexCoord4fARB)
USE_GL_FUNC(glMultiTexCoord4fvARB)
USE_GL_FUNC(glMultiTexCoord4svARB)
/* GL_ARB_occlusion_query */
USE_GL_FUNC(glBeginQueryARB)
USE_GL_FUNC(glDeleteQueriesARB)
USE_GL_FUNC(glEndQueryARB)
USE_GL_FUNC(glGenQueriesARB)
USE_GL_FUNC(glGetQueryivARB)
USE_GL_FUNC(glGetQueryObjectivARB)
USE_GL_FUNC(glGetQueryObjectuivARB)
USE_GL_FUNC(glIsQueryARB)
/* GL_ARB_point_parameters */
USE_GL_FUNC(glPointParameterfARB)
USE_GL_FUNC(glPointParameterfvARB)
/* GL_ARB_polgyon_offset_clamp */
USE_GL_FUNC(glPolygonOffsetClamp)
/* GL_ARB_provoking_vertex */
USE_GL_FUNC(glProvokingVertex)
/* GL_ARB_sample_shading */
USE_GL_FUNC(glMinSampleShadingARB)
/* GL_ARB_sampler_objects */
USE_GL_FUNC(glGenSamplers)
USE_GL_FUNC(glDeleteSamplers)
USE_GL_FUNC(glIsSampler)
USE_GL_FUNC(glBindSampler)
USE_GL_FUNC(glSamplerParameteri)
USE_GL_FUNC(glSamplerParameterf)
USE_GL_FUNC(glSamplerParameteriv)
USE_GL_FUNC(glSamplerParameterfv)
USE_GL_FUNC(glSamplerParameterIiv)
USE_GL_FUNC(glSamplerParameterIuiv)
USE_GL_FUNC(glGetSamplerParameteriv)
USE_GL_FUNC(glGetSamplerParameterfv)
USE_GL_FUNC(glGetSamplerParameterIiv)
USE_GL_FUNC(glGetSamplerParameterIuiv)
/* GL_ARB_shader_atomic_counters */
USE_GL_FUNC(glGetActiveAtomicCounterBufferiv)
/* GL_ARB_shader_image_load_store */
USE_GL_FUNC(glBindImageTexture)
USE_GL_FUNC(glMemoryBarrier)
/* GL_ARB_shader_objects */
USE_GL_FUNC(glAttachObjectARB)
USE_GL_FUNC(glBindAttribLocationARB)
USE_GL_FUNC(glCompileShaderARB)
USE_GL_FUNC(glCreateProgramObjectARB)
USE_GL_FUNC(glCreateShaderObjectARB)
USE_GL_FUNC(glDeleteObjectARB)
USE_GL_FUNC(glDetachObjectARB)
USE_GL_FUNC(glGetActiveUniformARB)
USE_GL_FUNC(glGetAttachedObjectsARB)
USE_GL_FUNC(glGetAttribLocationARB)
USE_GL_FUNC(glGetHandleARB)
USE_GL_FUNC(glGetInfoLogARB)
USE_GL_FUNC(glGetObjectParameterfvARB)
USE_GL_FUNC(glGetObjectParameterivARB)
USE_GL_FUNC(glGetShaderSourceARB)
USE_GL_FUNC(glGetUniformLocationARB)
USE_GL_FUNC(glGetUniformfvARB)
USE_GL_FUNC(glGetUniformivARB)
USE_GL_FUNC(glLinkProgramARB)
USE_GL_FUNC(glShaderSourceARB)
USE_GL_FUNC(glUniform1fARB)
USE_GL_FUNC(glUniform1fvARB)
USE_GL_FUNC(glUniform1iARB)
USE_GL_FUNC(glUniform1ivARB)
USE_GL_FUNC(glUniform2fARB)
USE_GL_FUNC(glUniform2fvARB)
USE_GL_FUNC(glUniform2iARB)
USE_GL_FUNC(glUniform2ivARB)
USE_GL_FUNC(glUniform3fARB)
USE_GL_FUNC(glUniform3fvARB)
USE_GL_FUNC(glUniform3iARB)
USE_GL_FUNC(glUniform3ivARB)
USE_GL_FUNC(glUniform4fARB)
USE_GL_FUNC(glUniform4fvARB)
USE_GL_FUNC(glUniform4iARB)
USE_GL_FUNC(glUniform4ivARB)
USE_GL_FUNC(glUniformMatrix2fvARB)
USE_GL_FUNC(glUniformMatrix3fvARB)
USE_GL_FUNC(glUniformMatrix4fvARB)
USE_GL_FUNC(glUseProgramObjectARB)
USE_GL_FUNC(glValidateProgramARB)
/* GL_ARB_shader_storage_buffer_object */
USE_GL_FUNC(glShaderStorageBlockBinding)
/* GL_ARB_sync */
USE_GL_FUNC(glClientWaitSync)
USE_GL_FUNC(glDeleteSync)
USE_GL_FUNC(glFenceSync)
USE_GL_FUNC(glGetInteger64v)
USE_GL_FUNC(glGetSynciv)
USE_GL_FUNC(glIsSync)
USE_GL_FUNC(glWaitSync)
/* GL_ARB_tessellation_shader */
USE_GL_FUNC(glPatchParameteri)
USE_GL_FUNC(glPatchParameterfv)
/* GL_ARB_texture_buffer_object */
USE_GL_FUNC(glTexBufferARB)
/* GL_ARB_texture_buffer_range */
USE_GL_FUNC(glTexBufferRange)
/* GL_ARB_texture_compression */
USE_GL_FUNC(glCompressedTexImage2DARB)
USE_GL_FUNC(glCompressedTexImage3DARB)
USE_GL_FUNC(glCompressedTexSubImage2DARB)
USE_GL_FUNC(glCompressedTexSubImage3DARB)
USE_GL_FUNC(glGetCompressedTexImageARB)
/* GL_ARB_texture_multisample */
USE_GL_FUNC(glGetMultisamplefv);
USE_GL_FUNC(glSampleMaski);
USE_GL_FUNC(glTexImage2DMultisample);
USE_GL_FUNC(glTexImage3DMultisample);
/* GL_ARB_texture_storage */
USE_GL_FUNC(glTexStorage1D)
USE_GL_FUNC(glTexStorage2D)
USE_GL_FUNC(glTexStorage3D)
/* GL_ARB_texture_storage_multisample */
USE_GL_FUNC(glTexStorage2DMultisample);
USE_GL_FUNC(glTexStorage3DMultisample);
/* GL_ARB_texture_view */
USE_GL_FUNC(glTextureView)
/* GL_ARB_timer_query */
USE_GL_FUNC(glQueryCounter)
USE_GL_FUNC(glGetQueryObjectui64v)
/* GL_ARB_transform_feedback2 */
USE_GL_FUNC(glBindTransformFeedback);
USE_GL_FUNC(glDeleteTransformFeedbacks);
USE_GL_FUNC(glDrawTransformFeedback);
USE_GL_FUNC(glGenTransformFeedbacks);
USE_GL_FUNC(glIsTransformFeedback);
USE_GL_FUNC(glPauseTransformFeedback);
USE_GL_FUNC(glResumeTransformFeedback);
/* GL_ARB_transform_feedback3 */
USE_GL_FUNC(glBeginQueryIndexed);
USE_GL_FUNC(glDrawTransformFeedbackStream);
USE_GL_FUNC(glEndQueryIndexed);
USE_GL_FUNC(glGetQueryIndexediv);
/* GL_ARB_uniform_buffer_object */
USE_GL_FUNC(glBindBufferBase)
USE_GL_FUNC(glBindBufferRange)
USE_GL_FUNC(glGetActiveUniformBlockName)
USE_GL_FUNC(glGetActiveUniformBlockiv)
USE_GL_FUNC(glGetActiveUniformName)
USE_GL_FUNC(glGetActiveUniformsiv)
USE_GL_FUNC(glGetIntegeri_v)
USE_GL_FUNC(glGetUniformBlockIndex)
USE_GL_FUNC(glGetUniformIndices)
USE_GL_FUNC(glUniformBlockBinding)
/* GL_ARB_vertex_buffer_object */
USE_GL_FUNC(glBindBufferARB)
USE_GL_FUNC(glBufferDataARB)
USE_GL_FUNC(glBufferSubDataARB)
USE_GL_FUNC(glDeleteBuffersARB)
USE_GL_FUNC(glGenBuffersARB)
USE_GL_FUNC(glGetBufferParameterivARB)
USE_GL_FUNC(glGetBufferPointervARB)
USE_GL_FUNC(glGetBufferSubDataARB)
USE_GL_FUNC(glIsBufferARB)
USE_GL_FUNC(glMapBufferARB)
USE_GL_FUNC(glUnmapBufferARB)
/* GL_ARB_vertex_program */
USE_GL_FUNC(glBindProgramARB)
USE_GL_FUNC(glDeleteProgramsARB)
USE_GL_FUNC(glDisableVertexAttribArrayARB)
USE_GL_FUNC(glEnableVertexAttribArrayARB)
USE_GL_FUNC(glGenProgramsARB)
USE_GL_FUNC(glGetProgramivARB)
USE_GL_FUNC(glProgramEnvParameter4fvARB)
USE_GL_FUNC(glProgramLocalParameter4fvARB)
USE_GL_FUNC(glProgramStringARB)
USE_GL_FUNC(glVertexAttrib1dARB)
USE_GL_FUNC(glVertexAttrib1dvARB)
USE_GL_FUNC(glVertexAttrib1fARB)
USE_GL_FUNC(glVertexAttrib1fvARB)
USE_GL_FUNC(glVertexAttrib1sARB)
USE_GL_FUNC(glVertexAttrib1svARB)
USE_GL_FUNC(glVertexAttrib2dARB)
USE_GL_FUNC(glVertexAttrib2dvARB)
USE_GL_FUNC(glVertexAttrib2fARB)
USE_GL_FUNC(glVertexAttrib2fvARB)
USE_GL_FUNC(glVertexAttrib2sARB)
USE_GL_FUNC(glVertexAttrib2svARB)
USE_GL_FUNC(glVertexAttrib3dARB)
USE_GL_FUNC(glVertexAttrib3dvARB)
USE_GL_FUNC(glVertexAttrib3fARB)
USE_GL_FUNC(glVertexAttrib3fvARB)
USE_GL_FUNC(glVertexAttrib3sARB)
USE_GL_FUNC(glVertexAttrib3svARB)
USE_GL_FUNC(glVertexAttrib4NbvARB)
USE_GL_FUNC(glVertexAttrib4NivARB)
USE_GL_FUNC(glVertexAttrib4NsvARB)
USE_GL_FUNC(glVertexAttrib4NubARB)
USE_GL_FUNC(glVertexAttrib4NubvARB)
USE_GL_FUNC(glVertexAttrib4NuivARB)
USE_GL_FUNC(glVertexAttrib4NusvARB)
USE_GL_FUNC(glVertexAttrib4bvARB)
USE_GL_FUNC(glVertexAttrib4dARB)
USE_GL_FUNC(glVertexAttrib4dvARB)
USE_GL_FUNC(glVertexAttrib4fARB)
USE_GL_FUNC(glVertexAttrib4fvARB)
USE_GL_FUNC(glVertexAttrib4ivARB)
USE_GL_FUNC(glVertexAttrib4sARB)
USE_GL_FUNC(glVertexAttrib4svARB)
USE_GL_FUNC(glVertexAttrib4ubvARB)
USE_GL_FUNC(glVertexAttrib4uivARB)
USE_GL_FUNC(glVertexAttrib4usvARB)
USE_GL_FUNC(glVertexAttribPointerARB)
/* GL_ARB_viewport_array */
USE_GL_FUNC(glDepthRangeArrayv)
USE_GL_FUNC(glDepthRangeIndexed)
USE_GL_FUNC(glGetDoublei_v)
USE_GL_FUNC(glGetFloati_v)
USE_GL_FUNC(glScissorArrayv)
USE_GL_FUNC(glScissorIndexed)
USE_GL_FUNC(glScissorIndexedv)
USE_GL_FUNC(glViewportArrayv)
USE_GL_FUNC(glViewportIndexedf)
USE_GL_FUNC(glViewportIndexedfv)
/* GL_ATI_fragment_shader */
USE_GL_FUNC(glAlphaFragmentOp1ATI)
USE_GL_FUNC(glAlphaFragmentOp2ATI)
USE_GL_FUNC(glAlphaFragmentOp3ATI)
USE_GL_FUNC(glBeginFragmentShaderATI)
USE_GL_FUNC(glBindFragmentShaderATI)
USE_GL_FUNC(glColorFragmentOp1ATI)
USE_GL_FUNC(glColorFragmentOp2ATI)
USE_GL_FUNC(glColorFragmentOp3ATI)
USE_GL_FUNC(glDeleteFragmentShaderATI)
USE_GL_FUNC(glEndFragmentShaderATI)
USE_GL_FUNC(glGenFragmentShadersATI)
USE_GL_FUNC(glPassTexCoordATI)
USE_GL_FUNC(glSampleMapATI)
USE_GL_FUNC(glSetFragmentShaderConstantATI)
/* GL_ATI_separate_stencil */
USE_GL_FUNC(glStencilOpSeparateATI)
USE_GL_FUNC(glStencilFuncSeparateATI)
/* GL_EXT_blend_color */
USE_GL_FUNC(glBlendColorEXT)
/* GL_EXT_blend_equation_separate */
USE_GL_FUNC(glBlendFuncSeparateEXT)
/* GL_EXT_blend_func_separate */
USE_GL_FUNC(glBlendEquationSeparateEXT)
/* GL_EXT_blend_minmax */
USE_GL_FUNC(glBlendEquationEXT)
/* GL_EXT_depth_bounds_test */
USE_GL_FUNC(glDepthBoundsEXT)
/* GL_EXT_draw_buffers2 */
USE_GL_FUNC(glColorMaskIndexedEXT)
USE_GL_FUNC(glDisableIndexedEXT)
USE_GL_FUNC(glEnableIndexedEXT)
USE_GL_FUNC(glGetBooleanIndexedvEXT)
USE_GL_FUNC(glGetIntegerIndexedvEXT)
USE_GL_FUNC(glIsEnabledIndexedEXT)
/* GL_EXT_fog_coord */
USE_GL_FUNC(glFogCoordPointerEXT)
USE_GL_FUNC(glFogCoorddEXT)
USE_GL_FUNC(glFogCoorddvEXT)
USE_GL_FUNC(glFogCoordfEXT)
USE_GL_FUNC(glFogCoordfvEXT)
/* GL_EXT_framebuffer_blit */
USE_GL_FUNC(glBlitFramebufferEXT)
/* GL_EXT_framebuffer_multisample */
USE_GL_FUNC(glRenderbufferStorageMultisampleEXT)
/* GL_EXT_framebuffer_object */
USE_GL_FUNC(glBindFramebufferEXT)
USE_GL_FUNC(glBindRenderbufferEXT)
USE_GL_FUNC(glCheckFramebufferStatusEXT)
USE_GL_FUNC(glDeleteFramebuffersEXT)
USE_GL_FUNC(glDeleteRenderbuffersEXT)
USE_GL_FUNC(glFramebufferRenderbufferEXT)
USE_GL_FUNC(glFramebufferTexture1DEXT)
USE_GL_FUNC(glFramebufferTexture2DEXT)
USE_GL_FUNC(glFramebufferTexture3DEXT)
USE_GL_FUNC(glGenFramebuffersEXT)
USE_GL_FUNC(glGenRenderbuffersEXT)
USE_GL_FUNC(glGenerateMipmapEXT)
USE_GL_FUNC(glGetFramebufferAttachmentParameterivEXT)
USE_GL_FUNC(glGetRenderbufferParameterivEXT)
USE_GL_FUNC(glIsFramebufferEXT)
USE_GL_FUNC(glIsRenderbufferEXT)
USE_GL_FUNC(glRenderbufferStorageEXT)
/* GL_EXT_gpu_program_parameters */
USE_GL_FUNC(glProgramEnvParameters4fvEXT)
USE_GL_FUNC(glProgramLocalParameters4fvEXT)
/* GL_EXT_gpu_shader4 */
USE_GL_FUNC(glBindFragDataLocationEXT)
USE_GL_FUNC(glGetFragDataLocationEXT)
USE_GL_FUNC(glGetUniformuivEXT)
USE_GL_FUNC(glGetVertexAttribIivEXT)
USE_GL_FUNC(glGetVertexAttribIuivEXT)
USE_GL_FUNC(glUniform1uiEXT)
USE_GL_FUNC(glUniform1uivEXT)
USE_GL_FUNC(glUniform2uiEXT)
USE_GL_FUNC(glUniform2uivEXT)
USE_GL_FUNC(glUniform3uiEXT)
USE_GL_FUNC(glUniform3uivEXT)
USE_GL_FUNC(glUniform4uiEXT)
USE_GL_FUNC(glUniform4uivEXT)
USE_GL_FUNC(glVertexAttribI1iEXT)
USE_GL_FUNC(glVertexAttribI1ivEXT)
USE_GL_FUNC(glVertexAttribI1uiEXT)
USE_GL_FUNC(glVertexAttribI1uivEXT)
USE_GL_FUNC(glVertexAttribI2iEXT)
USE_GL_FUNC(glVertexAttribI2ivEXT)
USE_GL_FUNC(glVertexAttribI2uiEXT)
USE_GL_FUNC(glVertexAttribI2uivEXT)
USE_GL_FUNC(glVertexAttribI3iEXT)
USE_GL_FUNC(glVertexAttribI3ivEXT)
USE_GL_FUNC(glVertexAttribI3uiEXT)
USE_GL_FUNC(glVertexAttribI3uivEXT)
USE_GL_FUNC(glVertexAttribI4bvEXT)
USE_GL_FUNC(glVertexAttribI4iEXT)
USE_GL_FUNC(glVertexAttribI4ivEXT)
USE_GL_FUNC(glVertexAttribI4svEXT)
USE_GL_FUNC(glVertexAttribI4ubvEXT)
USE_GL_FUNC(glVertexAttribI4uiEXT)
USE_GL_FUNC(glVertexAttribI4uivEXT)
USE_GL_FUNC(glVertexAttribI4usvEXT)
USE_GL_FUNC(glVertexAttribIPointerEXT)
/* GL_EXT_memory_object */
USE_GL_FUNC(glGetUnsignedBytei_vEXT)
USE_GL_FUNC(glGetUnsignedBytevEXT)
/* GL_EXT_point_parameters */
USE_GL_FUNC(glPointParameterfEXT)
USE_GL_FUNC(glPointParameterfvEXT)
/* GL_EXT_polgyon_offset_clamp */
USE_GL_FUNC(glPolygonOffsetClampEXT)
/* GL_EXT_provoking_vertex */
USE_GL_FUNC(glProvokingVertexEXT)
/* GL_EXT_secondary_color */
USE_GL_FUNC(glSecondaryColor3fEXT)
USE_GL_FUNC(glSecondaryColor3fvEXT)
USE_GL_FUNC(glSecondaryColor3ubEXT)
USE_GL_FUNC(glSecondaryColor3ubvEXT)
USE_GL_FUNC(glSecondaryColorPointerEXT)
/* GL_EXT_stencil_two_side */
USE_GL_FUNC(glActiveStencilFaceEXT)
/* GL_EXT_texture3D */
USE_GL_FUNC(glTexImage3D)
USE_GL_FUNC(glTexImage3DEXT)
USE_GL_FUNC(glTexSubImage3D)
USE_GL_FUNC(glTexSubImage3DEXT)
/* GL_NV_fence */
USE_GL_FUNC(glDeleteFencesNV)
USE_GL_FUNC(glFinishFenceNV)
USE_GL_FUNC(glGenFencesNV)
USE_GL_FUNC(glGetFenceivNV)
USE_GL_FUNC(glIsFenceNV)
USE_GL_FUNC(glSetFenceNV)
USE_GL_FUNC(glTestFenceNV)
/* GL_NV_half_float */
USE_GL_FUNC(glColor3hNV)
USE_GL_FUNC(glColor3hvNV)
USE_GL_FUNC(glColor4hNV)
USE_GL_FUNC(glColor4hvNV)
USE_GL_FUNC(glFogCoordhNV)
USE_GL_FUNC(glFogCoordhvNV)
USE_GL_FUNC(glMultiTexCoord1hNV)
USE_GL_FUNC(glMultiTexCoord1hvNV)
USE_GL_FUNC(glMultiTexCoord2hNV)
USE_GL_FUNC(glMultiTexCoord2hvNV)
USE_GL_FUNC(glMultiTexCoord3hNV)
USE_GL_FUNC(glMultiTexCoord3hvNV)
USE_GL_FUNC(glMultiTexCoord4hNV)
USE_GL_FUNC(glMultiTexCoord4hvNV)
USE_GL_FUNC(glNormal3hNV)
USE_GL_FUNC(glNormal3hvNV)
USE_GL_FUNC(glSecondaryColor3hNV)
USE_GL_FUNC(glSecondaryColor3hvNV)
USE_GL_FUNC(glTexCoord1hNV)
USE_GL_FUNC(glTexCoord1hvNV)
USE_GL_FUNC(glTexCoord2hNV)
USE_GL_FUNC(glTexCoord2hvNV)
USE_GL_FUNC(glTexCoord3hNV)
USE_GL_FUNC(glTexCoord3hvNV)
USE_GL_FUNC(glTexCoord4hNV)
USE_GL_FUNC(glTexCoord4hvNV)
USE_GL_FUNC(glVertex2hNV)
USE_GL_FUNC(glVertex2hvNV)
USE_GL_FUNC(glVertex3hNV)
USE_GL_FUNC(glVertex3hvNV)
USE_GL_FUNC(glVertex4hNV)
USE_GL_FUNC(glVertex4hvNV)
USE_GL_FUNC(glVertexAttrib1hNV)
USE_GL_FUNC(glVertexAttrib1hvNV)
USE_GL_FUNC(glVertexAttrib2hNV)
USE_GL_FUNC(glVertexAttrib2hvNV)
USE_GL_FUNC(glVertexAttrib3hNV)
USE_GL_FUNC(glVertexAttrib3hvNV)
USE_GL_FUNC(glVertexAttrib4hNV)
USE_GL_FUNC(glVertexAttrib4hvNV)
USE_GL_FUNC(glVertexAttribs1hvNV)
USE_GL_FUNC(glVertexAttribs2hvNV)
USE_GL_FUNC(glVertexAttribs3hvNV)
USE_GL_FUNC(glVertexAttribs4hvNV)
USE_GL_FUNC(glVertexWeighthNV)
USE_GL_FUNC(glVertexWeighthvNV)
/* GL_NV_point_sprite */
USE_GL_FUNC(glPointParameteriNV)
USE_GL_FUNC(glPointParameterivNV)
/* GL_NV_register_combiners */
USE_GL_FUNC(glCombinerInputNV)
USE_GL_FUNC(glCombinerOutputNV)
USE_GL_FUNC(glCombinerParameterfNV)
USE_GL_FUNC(glCombinerParameterfvNV)
USE_GL_FUNC(glCombinerParameteriNV)
USE_GL_FUNC(glCombinerParameterivNV)
USE_GL_FUNC(glFinalCombinerInputNV)
/* WGL extensions */
USE_GL_FUNC(wglChoosePixelFormatARB)
USE_GL_FUNC(wglGetExtensionsStringARB)
USE_GL_FUNC(wglGetPixelFormatAttribfvARB)
USE_GL_FUNC(wglGetPixelFormatAttribivARB)
USE_GL_FUNC(wglQueryCurrentRendererIntegerWINE)
USE_GL_FUNC(wglQueryCurrentRendererStringWINE)
USE_GL_FUNC(wglQueryRendererIntegerWINE)
USE_GL_FUNC(wglQueryRendererStringWINE)
USE_GL_FUNC(wglSetPixelFormatWINE)
USE_GL_FUNC(wglSwapIntervalEXT)
/* Newer core functions */
USE_GL_FUNC(glActiveTexture) /* OpenGL 1.3 */
USE_GL_FUNC(glAttachShader) /* OpenGL 2.0 */
USE_GL_FUNC(glBeginQuery) /* OpenGL 1.5 */
USE_GL_FUNC(glBeginTransformFeedback) /* OpenGL 3.0 */
USE_GL_FUNC(glBindAttribLocation) /* OpenGL 2.0 */
USE_GL_FUNC(glBindBuffer) /* OpenGL 1.5 */
USE_GL_FUNC(glBindFragDataLocation) /* OpenGL 3.0 */
USE_GL_FUNC(glBindVertexArray) /* OpenGL 3.0 */
USE_GL_FUNC(glBlendColor) /* OpenGL 1.4 */
USE_GL_FUNC(glBlendEquation) /* OpenGL 1.4 */
USE_GL_FUNC(glBlendEquationSeparate) /* OpenGL 2.0 */
USE_GL_FUNC(glBlendFuncSeparate) /* OpenGL 1.4 */
USE_GL_FUNC(glBufferData) /* OpenGL 1.5 */
USE_GL_FUNC(glBufferSubData) /* OpenGL 1.5 */
USE_GL_FUNC(glColorMaski) /* OpenGL 3.0 */
USE_GL_FUNC(glCompileShader) /* OpenGL 2.0 */
USE_GL_FUNC(glCompressedTexImage2D) /* OpenGL 1.3 */
USE_GL_FUNC(glCompressedTexImage3D) /* OpenGL 1.3 */
USE_GL_FUNC(glCompressedTexSubImage2D) /* OpenGL 1.3 */
USE_GL_FUNC(glCompressedTexSubImage3D) /* OpenGL 1.3 */
USE_GL_FUNC(glCreateProgram) /* OpenGL 2.0 */
USE_GL_FUNC(glCreateShader) /* OpenGL 2.0 */
USE_GL_FUNC(glDebugMessageCallback) /* OpenGL 4.3 */
USE_GL_FUNC(glDebugMessageControl) /* OpenGL 4.3 */
USE_GL_FUNC(glDebugMessageInsert) /* OpenGL 4.3 */
USE_GL_FUNC(glDeleteBuffers) /* OpenGL 1.5 */
USE_GL_FUNC(glDeleteProgram) /* OpenGL 2.0 */
USE_GL_FUNC(glDeleteQueries) /* OpenGL 1.5 */
USE_GL_FUNC(glDeleteShader) /* OpenGL 2.0 */
USE_GL_FUNC(glDeleteVertexArrays) /* OpenGL 3.0 */
USE_GL_FUNC(glDetachShader) /* OpenGL 2.0 */
USE_GL_FUNC(glDisablei) /* OpenGL 3.0 */
USE_GL_FUNC(glDisableVertexAttribArray) /* OpenGL 2.0 */
USE_GL_FUNC(glDrawArraysInstanced) /* OpenGL 3.1 */
USE_GL_FUNC(glDrawBuffers) /* OpenGL 2.0 */
USE_GL_FUNC(glDrawElementsInstanced) /* OpenGL 3.1 */
USE_GL_FUNC(glEnablei) /* OpenGL 3.0 */
USE_GL_FUNC(glEnableVertexAttribArray) /* OpenGL 2.0 */
USE_GL_FUNC(glEndQuery) /* OpenGL 1.5 */
USE_GL_FUNC(glEndTransformFeedback) /* OpenGL 3.0 */
USE_GL_FUNC(glFramebufferTexture) /* OpenGL 3.2 */
USE_GL_FUNC(glGenBuffers) /* OpenGL 1.5 */
USE_GL_FUNC(glGenQueries) /* OpenGL 1.5 */
USE_GL_FUNC(glGenVertexArrays) /* OpenGL 3.0 */
USE_GL_FUNC(glGetActiveUniform) /* OpenGL 2.0 */
USE_GL_FUNC(glGetAttachedShaders) /* OpenGL 2.0 */
USE_GL_FUNC(glGetAttribLocation) /* OpenGL 2.0 */
USE_GL_FUNC(glGetBooleani_v) /* OpenGL 3.0 */
USE_GL_FUNC(glGetBufferSubData) /* OpenGL 1.5 */
USE_GL_FUNC(glGetCompressedTexImage) /* OpenGL 1.3 */
USE_GL_FUNC(glGetDebugMessageLog) /* OpenGL 4.3 */
USE_GL_FUNC(glGetIntegeri_v) /* OpenGL 3.0 */
USE_GL_FUNC(glGetProgramInfoLog) /* OpenGL 2.0 */
USE_GL_FUNC(glGetProgramiv) /* OpenGL 2.0 */
USE_GL_FUNC(glGetQueryiv) /* OpenGL 1.5 */
USE_GL_FUNC(glGetQueryObjectuiv) /* OpenGL 1.5 */
USE_GL_FUNC(glGetShaderInfoLog) /* OpenGL 2.0 */
USE_GL_FUNC(glGetShaderiv) /* OpenGL 2.0 */
USE_GL_FUNC(glGetShaderSource) /* OpenGL 2.0 */
USE_GL_FUNC(glGetStringi) /* OpenGL 3.0 */
USE_GL_FUNC(glGetTextureLevelParameteriv) /* OpenGL 4.5 */
USE_GL_FUNC(glGetTextureParameteriv) /* OpenGL 4.5 */
USE_GL_FUNC(glGetUniformfv) /* OpenGL 2.0 */
USE_GL_FUNC(glGetUniformiv) /* OpenGL 2.0 */
USE_GL_FUNC(glGetUniformLocation) /* OpenGL 2.0 */
USE_GL_FUNC(glIsEnabledi) /* OpenGL 3.0 */
USE_GL_FUNC(glLinkProgram) /* OpenGL 2.0 */
USE_GL_FUNC(glMapBuffer) /* OpenGL 1.5 */
USE_GL_FUNC(glMinSampleShading) /* OpenGL 4.0 */
USE_GL_FUNC(glPointParameteri) /* OpenGL 1.4 */
USE_GL_FUNC(glPointParameteriv) /* OpenGL 1.4 */
USE_GL_FUNC(glShaderSource) /* OpenGL 2.0 */
USE_GL_FUNC(glStencilFuncSeparate) /* OpenGL 2.0 */
USE_GL_FUNC(glStencilOpSeparate) /* OpenGL 2.0 */
USE_GL_FUNC(glTexBuffer) /* OpenGL 3.1 */
USE_GL_FUNC(glTexImage3D) /* OpenGL 1.2 */
USE_GL_FUNC(glTexSubImage3D) /* OpenGL 1.2 */
USE_GL_FUNC(glTransformFeedbackVaryings) /* OpenGL 3.0 */
USE_GL_FUNC(glUniform1f) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform1fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform1i) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform1iv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform2f) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform2fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform2i) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform2iv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform3f) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform3fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform3i) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform3iv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform4f) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform4fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform4i) /* OpenGL 2.0 */
USE_GL_FUNC(glUniform4iv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniformMatrix2fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniformMatrix3fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUniformMatrix4fv) /* OpenGL 2.0 */
USE_GL_FUNC(glUnmapBuffer) /* OpenGL 1.5 */
USE_GL_FUNC(glUseProgram) /* OpenGL 2.0 */
USE_GL_FUNC(glValidateProgram) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib1f) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib1fv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib2f) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib2fv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib3f) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib3fv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4f) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4fv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4Nsv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4Nub) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4Nubv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4Nusv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4sv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttrib4ubv) /* OpenGL 2.0 */
USE_GL_FUNC(glVertexAttribDivisor) /* OpenGL 3.3 */
USE_GL_FUNC(glVertexAttribIPointer) /* OpenGL 3.0 */
USE_GL_FUNC(glVertexAttribPointer) /* OpenGL 2.0 */
#undef USE_GL_FUNC
#ifndef USE_WIN32_OPENGL
/* hack: use the functions directly from the TEB table to bypass the thunks */
/* note that we still need the above wglGetProcAddress calls to initialize the table */
gl_info->gl_ops.ext = ((struct opengl_funcs *)NtCurrentTeb()->glTable)->ext;
#endif
#define MAP_GL_FUNCTION(core_func, ext_func) \
do \
{ \
if (!gl_info->gl_ops.ext.p_##core_func) \
gl_info->gl_ops.ext.p_##core_func = gl_info->gl_ops.ext.p_##ext_func; \
} while (0)
#define MAP_GL_FUNCTION_CAST(core_func, ext_func) \
do \
{ \
if (!gl_info->gl_ops.ext.p_##core_func) \
gl_info->gl_ops.ext.p_##core_func = (void *)gl_info->gl_ops.ext.p_##ext_func; \
} while (0)
MAP_GL_FUNCTION(glActiveTexture, glActiveTextureARB);
MAP_GL_FUNCTION(glAttachShader, glAttachObjectARB);
MAP_GL_FUNCTION(glBeginQuery, glBeginQueryARB);
MAP_GL_FUNCTION(glBindAttribLocation, glBindAttribLocationARB);
MAP_GL_FUNCTION(glBindBuffer, glBindBufferARB);
MAP_GL_FUNCTION(glBindFragDataLocation, glBindFragDataLocationEXT);
MAP_GL_FUNCTION(glBlendColor, glBlendColorEXT);
MAP_GL_FUNCTION(glBlendEquation, glBlendEquationEXT);
MAP_GL_FUNCTION(glBlendEquationSeparate, glBlendEquationSeparateEXT);
MAP_GL_FUNCTION(glBlendFuncSeparate, glBlendFuncSeparateEXT);
MAP_GL_FUNCTION(glBufferData, glBufferDataARB);
MAP_GL_FUNCTION(glBufferSubData, glBufferSubDataARB);
MAP_GL_FUNCTION(glColorMaski, glColorMaskIndexedEXT);
MAP_GL_FUNCTION(glCompileShader, glCompileShaderARB);
MAP_GL_FUNCTION(glCompressedTexImage2D, glCompressedTexImage2DARB);
MAP_GL_FUNCTION(glCompressedTexImage3D, glCompressedTexImage3DARB);
MAP_GL_FUNCTION(glCompressedTexSubImage2D, glCompressedTexSubImage2DARB);
MAP_GL_FUNCTION(glCompressedTexSubImage3D, glCompressedTexSubImage3DARB);
MAP_GL_FUNCTION(glCreateProgram, glCreateProgramObjectARB);
MAP_GL_FUNCTION(glCreateShader, glCreateShaderObjectARB);
MAP_GL_FUNCTION(glDebugMessageCallback, glDebugMessageCallbackARB);
MAP_GL_FUNCTION(glDebugMessageControl, glDebugMessageControlARB);
MAP_GL_FUNCTION(glDebugMessageInsert, glDebugMessageInsertARB);
MAP_GL_FUNCTION(glDeleteBuffers, glDeleteBuffersARB);
MAP_GL_FUNCTION(glDeleteProgram, glDeleteObjectARB);
MAP_GL_FUNCTION(glDeleteQueries, glDeleteQueriesARB);
MAP_GL_FUNCTION(glDeleteShader, glDeleteObjectARB);
MAP_GL_FUNCTION(glDetachShader, glDetachObjectARB);
MAP_GL_FUNCTION(glDisablei, glDisableIndexedEXT);
MAP_GL_FUNCTION(glDisableVertexAttribArray, glDisableVertexAttribArrayARB);
MAP_GL_FUNCTION(glDrawArraysInstanced, glDrawArraysInstancedARB);
MAP_GL_FUNCTION(glDrawBuffers, glDrawBuffersARB);
MAP_GL_FUNCTION(glDrawElementsInstanced, glDrawElementsInstancedARB);
MAP_GL_FUNCTION(glEnablei, glEnableIndexedEXT);
MAP_GL_FUNCTION(glEnableVertexAttribArray, glEnableVertexAttribArrayARB);
MAP_GL_FUNCTION(glEndQuery, glEndQueryARB);
MAP_GL_FUNCTION(glFramebufferTexture, glFramebufferTextureARB);
MAP_GL_FUNCTION(glGenBuffers, glGenBuffersARB);
MAP_GL_FUNCTION(glGenQueries, glGenQueriesARB);
MAP_GL_FUNCTION(glGetActiveUniform, glGetActiveUniformARB);
MAP_GL_FUNCTION(glGetAttachedShaders, glGetAttachedObjectsARB);
MAP_GL_FUNCTION(glGetAttribLocation, glGetAttribLocationARB);
MAP_GL_FUNCTION(glGetBooleani_v, glGetBooleanIndexedvEXT);
MAP_GL_FUNCTION(glGetBufferSubData, glGetBufferSubDataARB);
MAP_GL_FUNCTION(glGetCompressedTexImage, glGetCompressedTexImageARB);
MAP_GL_FUNCTION(glGetDebugMessageLog, glGetDebugMessageLogARB);
MAP_GL_FUNCTION(glGetIntegeri_v, glGetIntegerIndexedvEXT);
MAP_GL_FUNCTION(glGetProgramInfoLog, glGetInfoLogARB);
MAP_GL_FUNCTION(glGetProgramiv, glGetObjectParameterivARB);
MAP_GL_FUNCTION(glGetQueryiv, glGetQueryivARB);
MAP_GL_FUNCTION(glGetQueryObjectuiv, glGetQueryObjectuivARB);
MAP_GL_FUNCTION(glGetShaderInfoLog, glGetInfoLogARB);
MAP_GL_FUNCTION(glGetShaderiv, glGetObjectParameterivARB);
MAP_GL_FUNCTION(glGetShaderSource, glGetShaderSourceARB);
MAP_GL_FUNCTION(glGetUniformfv, glGetUniformfvARB);
MAP_GL_FUNCTION(glGetUniformiv, glGetUniformivARB);
MAP_GL_FUNCTION(glGetUniformLocation, glGetUniformLocationARB);
MAP_GL_FUNCTION(glIsEnabledi, glIsEnabledIndexedEXT);
MAP_GL_FUNCTION(glLinkProgram, glLinkProgramARB);
MAP_GL_FUNCTION(glMapBuffer, glMapBufferARB);
MAP_GL_FUNCTION(glMinSampleShading, glMinSampleShadingARB);
MAP_GL_FUNCTION(glPolygonOffsetClamp, glPolygonOffsetClampEXT);
MAP_GL_FUNCTION_CAST(glShaderSource, glShaderSourceARB);
MAP_GL_FUNCTION(glTexBuffer, glTexBufferARB);
MAP_GL_FUNCTION_CAST(glTexImage3D, glTexImage3DEXT);
MAP_GL_FUNCTION(glTexSubImage3D, glTexSubImage3DEXT);
MAP_GL_FUNCTION(glUniform1f, glUniform1fARB);
MAP_GL_FUNCTION(glUniform1fv, glUniform1fvARB);
MAP_GL_FUNCTION(glUniform1i, glUniform1iARB);
MAP_GL_FUNCTION(glUniform1iv, glUniform1ivARB);
MAP_GL_FUNCTION(glUniform2f, glUniform2fARB);
MAP_GL_FUNCTION(glUniform2fv, glUniform2fvARB);
MAP_GL_FUNCTION(glUniform2i, glUniform2iARB);
MAP_GL_FUNCTION(glUniform2iv, glUniform2ivARB);
MAP_GL_FUNCTION(glUniform3f, glUniform3fARB);
MAP_GL_FUNCTION(glUniform3fv, glUniform3fvARB);
MAP_GL_FUNCTION(glUniform3i, glUniform3iARB);
MAP_GL_FUNCTION(glUniform3iv, glUniform3ivARB);
MAP_GL_FUNCTION(glUniform4f, glUniform4fARB);
MAP_GL_FUNCTION(glUniform4fv, glUniform4fvARB);
MAP_GL_FUNCTION(glUniform4i, glUniform4iARB);
MAP_GL_FUNCTION(glUniform4iv, glUniform4ivARB);
MAP_GL_FUNCTION(glUniformMatrix2fv, glUniformMatrix2fvARB);
MAP_GL_FUNCTION(glUniformMatrix3fv, glUniformMatrix3fvARB);
MAP_GL_FUNCTION(glUniformMatrix4fv, glUniformMatrix4fvARB);
MAP_GL_FUNCTION(glUnmapBuffer, glUnmapBufferARB);
MAP_GL_FUNCTION(glUseProgram, glUseProgramObjectARB);
MAP_GL_FUNCTION(glValidateProgram, glValidateProgramARB);
MAP_GL_FUNCTION(glVertexAttrib1f, glVertexAttrib1fARB);
MAP_GL_FUNCTION(glVertexAttrib1fv, glVertexAttrib1fvARB);
MAP_GL_FUNCTION(glVertexAttrib2f, glVertexAttrib2fARB);
MAP_GL_FUNCTION(glVertexAttrib2fv, glVertexAttrib2fvARB);
MAP_GL_FUNCTION(glVertexAttrib3f, glVertexAttrib3fARB);
MAP_GL_FUNCTION(glVertexAttrib3fv, glVertexAttrib3fvARB);
MAP_GL_FUNCTION(glVertexAttrib4f, glVertexAttrib4fARB);
MAP_GL_FUNCTION(glVertexAttrib4fv, glVertexAttrib4fvARB);
MAP_GL_FUNCTION(glVertexAttrib4Nsv, glVertexAttrib4NsvARB);
MAP_GL_FUNCTION(glVertexAttrib4Nub, glVertexAttrib4NubARB);
MAP_GL_FUNCTION(glVertexAttrib4Nubv, glVertexAttrib4NubvARB);
MAP_GL_FUNCTION(glVertexAttrib4Nusv, glVertexAttrib4NusvARB);
MAP_GL_FUNCTION(glVertexAttrib4sv, glVertexAttrib4svARB);
MAP_GL_FUNCTION(glVertexAttrib4ubv, glVertexAttrib4ubvARB);
MAP_GL_FUNCTION(glVertexAttribDivisor, glVertexAttribDivisorARB);
MAP_GL_FUNCTION(glVertexAttribIPointer, glVertexAttribIPointerEXT);
MAP_GL_FUNCTION(glVertexAttribPointer, glVertexAttribPointerARB);
#undef MAP_GL_FUNCTION
#undef MAP_GL_FUNCTION_CAST
}
static void wined3d_adapter_init_limits(struct wined3d_gl_info *gl_info)
{
unsigned int i, sampler_count;
GLint gl_max;
gl_info->limits.buffers = 1;
gl_info->limits.textures = 0;
gl_info->limits.texture_coords = 0;
for (i = 0; i < WINED3D_SHADER_TYPE_COUNT; ++i)
{
gl_info->limits.uniform_blocks[i] = 0;
gl_info->limits.samplers[i] = 0;
}
gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL] = 1;
gl_info->limits.combined_samplers = gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL];
gl_info->limits.graphics_samplers = gl_info->limits.combined_samplers;
gl_info->limits.vertex_attribs = 16;
gl_info->limits.texture_buffer_offset_alignment = 1;
gl_info->limits.glsl_vs_float_constants = 0;
gl_info->limits.glsl_ps_float_constants = 0;
gl_info->limits.arb_vs_float_constants = 0;
gl_info->limits.arb_vs_native_constants = 0;
gl_info->limits.arb_vs_instructions = 0;
gl_info->limits.arb_vs_temps = 0;
gl_info->limits.arb_ps_float_constants = 0;
gl_info->limits.arb_ps_local_constants = 0;
gl_info->limits.arb_ps_instructions = 0;
gl_info->limits.arb_ps_temps = 0;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_CLIP_DISTANCES, &gl_max);
gl_info->limits.user_clip_distances = min(WINED3D_MAX_CLIP_DISTANCES, gl_max);
TRACE("Clip plane support - max planes %d.\n", gl_max);
if (gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_LIGHTS, &gl_max);
gl_info->limits.lights = gl_max;
TRACE("Light support - max lights %d.\n", gl_max);
}
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TEXTURE_SIZE, &gl_max);
gl_info->limits.texture_size = gl_max;
TRACE("Maximum texture size support - max texture size %d.\n", gl_max);
if (gl_info->supported[ARB_MAP_BUFFER_ALIGNMENT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MIN_MAP_BUFFER_ALIGNMENT, &gl_max);
TRACE("Minimum buffer map alignment: %d.\n", gl_max);
}
else
{
WARN_(d3d_perf)("Driver doesn't guarantee a minimum buffer map alignment.\n");
}
if (gl_info->supported[NV_REGISTER_COMBINERS])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_GENERAL_COMBINERS_NV, &gl_max);
gl_info->limits.general_combiners = gl_max;
TRACE("Max general combiners: %d.\n", gl_max);
}
if (gl_info->supported[ARB_DRAW_BUFFERS] && wined3d_settings.offscreen_rendering_mode == ORM_FBO)
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_DRAW_BUFFERS_ARB, &gl_max);
gl_info->limits.buffers = min(MAX_RENDER_TARGET_VIEWS, gl_max);
TRACE("Max draw buffers: %u.\n", gl_max);
}
if (gl_info->supported[ARB_MULTITEXTURE])
{
if (gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TEXTURE_UNITS_ARB, &gl_max);
gl_info->limits.textures = min(WINED3D_MAX_TEXTURES, gl_max);
TRACE("Max textures: %d.\n", gl_info->limits.textures);
if (gl_info->supported[ARB_FRAGMENT_PROGRAM])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TEXTURE_COORDS_ARB, &gl_max);
gl_info->limits.texture_coords = min(WINED3D_MAX_TEXTURES, gl_max);
}
else
{
gl_info->limits.texture_coords = gl_info->limits.textures;
}
TRACE("Max texture coords: %d.\n", gl_info->limits.texture_coords);
}
if (gl_info->supported[ARB_FRAGMENT_PROGRAM] || gl_info->supported[ARB_FRAGMENT_SHADER])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TEXTURE_IMAGE_UNITS, &gl_max);
gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL] = gl_max;
}
else
{
gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL] = gl_info->limits.textures;
}
TRACE("Max fragment samplers: %d.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL]);
if (gl_info->supported[ARB_VERTEX_SHADER])
{
unsigned int vertex_sampler_count;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VERTEX_TEXTURE_IMAGE_UNITS_ARB, &gl_max);
vertex_sampler_count = gl_info->limits.samplers[WINED3D_SHADER_TYPE_VERTEX] = gl_max;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_COMBINED_TEXTURE_IMAGE_UNITS_ARB, &gl_max);
gl_info->limits.combined_samplers = gl_max;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VERTEX_ATTRIBS_ARB, &gl_max);
gl_info->limits.vertex_attribs = gl_max;
/* Loading GLSL sampler uniforms is much simpler if we can assume
* that the sampler setup is known at shader link time. In a
* vertex shader + pixel shader combination this isn't an issue
* because then the sampler setup only depends on the two shaders.
* If a pixel shader is used with fixed-function vertex processing
* we're fine too because fixed-function vertex processing doesn't
* use any samplers. If fixed-function fragment processing is used
* we have to make sure that all vertex sampler setups are valid
* together with all possible fixed-function fragment processing
* setups. This is true if vsamplers + WINED3D_MAX_TEXTURES <= max_samplers.
* This is true on all Direct3D 9 cards that support vertex
* texture fetch (GeForce 6 and GeForce 7 cards). Direct3D 9
* Radeon cards do not support vertex texture fetch. Direct3D 10
* cards have 128 samplers, and Direct3D 9 is limited to 8
* fixed-function texture stages and 4 vertex samplers.
* Direct3D 10 does not have a fixed-function pipeline anymore.
*
* So this is just a check to check that our assumption holds
* true. If not, write a warning and reduce the number of vertex
* samplers or probably disable vertex texture fetch. */
if (vertex_sampler_count && gl_info->limits.combined_samplers < 12
&& WINED3D_MAX_TEXTURES + vertex_sampler_count > gl_info->limits.combined_samplers)
{
FIXME("OpenGL implementation supports %u vertex samplers and %u total samplers.\n",
vertex_sampler_count, gl_info->limits.combined_samplers);
FIXME("Expected vertex samplers + WINED3D_MAX_TEXTURES(=8) > combined_samplers.\n");
if (gl_info->limits.combined_samplers > WINED3D_MAX_TEXTURES)
vertex_sampler_count = gl_info->limits.combined_samplers - WINED3D_MAX_TEXTURES;
else
vertex_sampler_count = 0;
gl_info->limits.samplers[WINED3D_SHADER_TYPE_VERTEX] = vertex_sampler_count;
}
}
else
{
gl_info->limits.combined_samplers = gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL];
}
TRACE("Max vertex samplers: %u.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_VERTEX]);
TRACE("Max combined samplers: %u.\n", gl_info->limits.combined_samplers);
TRACE("Max vertex attributes: %u.\n", gl_info->limits.vertex_attribs);
}
else
{
gl_info->limits.textures = 1;
gl_info->limits.texture_coords = 1;
}
if (gl_info->supported[EXT_TEXTURE3D])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_3D_TEXTURE_SIZE_EXT, &gl_max);
gl_info->limits.texture3d_size = gl_max;
TRACE("Max texture3D size: %d.\n", gl_info->limits.texture3d_size);
}
if (gl_info->supported[ARB_TEXTURE_FILTER_ANISOTROPIC])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TEXTURE_MAX_ANISOTROPY, &gl_max);
gl_info->limits.anisotropy = gl_max;
TRACE("Max anisotropy: %d.\n", gl_info->limits.anisotropy);
}
if (gl_info->supported[ARB_FRAGMENT_PROGRAM])
{
GL_EXTCALL(glGetProgramivARB(GL_FRAGMENT_PROGRAM_ARB, GL_MAX_PROGRAM_ENV_PARAMETERS_ARB, &gl_max));
gl_info->limits.arb_ps_float_constants = gl_max;
TRACE("Max ARB_FRAGMENT_PROGRAM float constants: %d.\n", gl_info->limits.arb_ps_float_constants);
GL_EXTCALL(glGetProgramivARB(GL_FRAGMENT_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_PARAMETERS_ARB, &gl_max));
gl_info->limits.arb_ps_native_constants = gl_max;
TRACE("Max ARB_FRAGMENT_PROGRAM native float constants: %d.\n",
gl_info->limits.arb_ps_native_constants);
GL_EXTCALL(glGetProgramivARB(GL_FRAGMENT_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_TEMPORARIES_ARB, &gl_max));
gl_info->limits.arb_ps_temps = gl_max;
TRACE("Max ARB_FRAGMENT_PROGRAM native temporaries: %d.\n", gl_info->limits.arb_ps_temps);
GL_EXTCALL(glGetProgramivARB(GL_FRAGMENT_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_INSTRUCTIONS_ARB, &gl_max));
gl_info->limits.arb_ps_instructions = gl_max;
TRACE("Max ARB_FRAGMENT_PROGRAM native instructions: %d.\n", gl_info->limits.arb_ps_instructions);
GL_EXTCALL(glGetProgramivARB(GL_FRAGMENT_PROGRAM_ARB, GL_MAX_PROGRAM_LOCAL_PARAMETERS_ARB, &gl_max));
gl_info->limits.arb_ps_local_constants = gl_max;
TRACE("Max ARB_FRAGMENT_PROGRAM local parameters: %d.\n", gl_info->limits.arb_ps_instructions);
}
if (gl_info->supported[ARB_VERTEX_PROGRAM])
{
GL_EXTCALL(glGetProgramivARB(GL_VERTEX_PROGRAM_ARB, GL_MAX_PROGRAM_ENV_PARAMETERS_ARB, &gl_max));
gl_info->limits.arb_vs_float_constants = gl_max;
TRACE("Max ARB_VERTEX_PROGRAM float constants: %d.\n", gl_info->limits.arb_vs_float_constants);
GL_EXTCALL(glGetProgramivARB(GL_VERTEX_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_PARAMETERS_ARB, &gl_max));
gl_info->limits.arb_vs_native_constants = gl_max;
TRACE("Max ARB_VERTEX_PROGRAM native float constants: %d.\n",
gl_info->limits.arb_vs_native_constants);
GL_EXTCALL(glGetProgramivARB(GL_VERTEX_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_TEMPORARIES_ARB, &gl_max));
gl_info->limits.arb_vs_temps = gl_max;
TRACE("Max ARB_VERTEX_PROGRAM native temporaries: %d.\n", gl_info->limits.arb_vs_temps);
GL_EXTCALL(glGetProgramivARB(GL_VERTEX_PROGRAM_ARB, GL_MAX_PROGRAM_NATIVE_INSTRUCTIONS_ARB, &gl_max));
gl_info->limits.arb_vs_instructions = gl_max;
TRACE("Max ARB_VERTEX_PROGRAM native instructions: %d.\n", gl_info->limits.arb_vs_instructions);
}
if (gl_info->supported[ARB_VERTEX_SHADER])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VERTEX_UNIFORM_COMPONENTS_ARB, &gl_max);
gl_info->limits.glsl_vs_float_constants = gl_max / 4;
TRACE("Max ARB_VERTEX_SHADER float constants: %u.\n", gl_info->limits.glsl_vs_float_constants);
if (gl_info->supported[ARB_UNIFORM_BUFFER_OBJECT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VERTEX_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_VERTEX] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max vertex uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_VERTEX], gl_max);
}
}
if (gl_info->supported[ARB_TESSELLATION_SHADER])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TESS_CONTROL_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_HULL] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max hull uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_HULL], gl_max);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TESS_CONTROL_TEXTURE_IMAGE_UNITS, &gl_max);
gl_info->limits.samplers[WINED3D_SHADER_TYPE_HULL] = gl_max;
TRACE("Max hull samplers: %u.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_HULL]);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TESS_EVALUATION_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_DOMAIN] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max domain uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_DOMAIN], gl_max);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_TESS_EVALUATION_TEXTURE_IMAGE_UNITS, &gl_max);
gl_info->limits.samplers[WINED3D_SHADER_TYPE_DOMAIN] = gl_max;
TRACE("Max domain samplers: %u.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_DOMAIN]);
}
if (gl_info->supported[WINED3D_GL_VERSION_3_2] && gl_info->supported[ARB_UNIFORM_BUFFER_OBJECT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_GEOMETRY_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_GEOMETRY] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max geometry uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_GEOMETRY], gl_max);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_GEOMETRY_TEXTURE_IMAGE_UNITS, &gl_max);
gl_info->limits.samplers[WINED3D_SHADER_TYPE_GEOMETRY] = gl_max;
TRACE("Max geometry samplers: %u.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_GEOMETRY]);
}
if (gl_info->supported[ARB_FRAGMENT_SHADER])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAGMENT_UNIFORM_COMPONENTS_ARB, &gl_max);
gl_info->limits.glsl_ps_float_constants = gl_max / 4;
TRACE("Max ARB_FRAGMENT_SHADER float constants: %u.\n", gl_info->limits.glsl_ps_float_constants);
if (gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VARYING_FLOATS_ARB, &gl_max);
gl_info->limits.glsl_varyings = gl_max;
}
else
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAGMENT_INPUT_COMPONENTS, &gl_max);
gl_info->limits.glsl_varyings = gl_max - 4;
}
TRACE("Max GLSL varyings: %u (%u 4 component varyings).\n", gl_info->limits.glsl_varyings,
gl_info->limits.glsl_varyings / 4);
if (gl_info->supported[ARB_UNIFORM_BUFFER_OBJECT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAGMENT_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_PIXEL] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max fragment uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_PIXEL], gl_max);
}
}
if (gl_info->supported[ARB_COMPUTE_SHADER])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_COMPUTE_UNIFORM_BLOCKS, &gl_max);
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_COMPUTE] = min(gl_max, WINED3D_MAX_CBS);
TRACE("Max compute uniform blocks: %u (%d).\n",
gl_info->limits.uniform_blocks[WINED3D_SHADER_TYPE_COMPUTE], gl_max);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_COMPUTE_TEXTURE_IMAGE_UNITS, &gl_max);
gl_info->limits.samplers[WINED3D_SHADER_TYPE_COMPUTE] = gl_max;
TRACE("Max compute samplers: %u.\n", gl_info->limits.samplers[WINED3D_SHADER_TYPE_COMPUTE]);
}
if (gl_info->supported[ARB_UNIFORM_BUFFER_OBJECT])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_COMBINED_UNIFORM_BLOCKS, &gl_max);
TRACE("Max combined uniform blocks: %d.\n", gl_max);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_UNIFORM_BUFFER_BINDINGS, &gl_max);
TRACE("Max uniform buffer bindings: %d.\n", gl_max);
}
if (gl_info->supported[ARB_TEXTURE_BUFFER_RANGE])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_TEXTURE_BUFFER_OFFSET_ALIGNMENT, &gl_max);
gl_info->limits.texture_buffer_offset_alignment = gl_max;
TRACE("Minimum required texture buffer offset alignment %d.\n", gl_max);
}
if (gl_info->supported[ARB_SHADER_ATOMIC_COUNTERS])
{
GLint max_fragment_buffers, max_combined_buffers, max_bindings;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAGMENT_ATOMIC_COUNTER_BUFFERS, &max_fragment_buffers);
TRACE("Max fragment atomic counter buffers: %d.\n", max_fragment_buffers);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_COMBINED_ATOMIC_COUNTER_BUFFERS, &max_combined_buffers);
TRACE("Max combined atomic counter buffers: %d.\n", max_combined_buffers);
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_ATOMIC_COUNTER_BUFFER_BINDINGS, &max_bindings);
TRACE("Max atomic counter buffer bindings: %d.\n", max_bindings);
if (max_fragment_buffers < MAX_UNORDERED_ACCESS_VIEWS
|| max_combined_buffers < MAX_UNORDERED_ACCESS_VIEWS
|| max_bindings < MAX_UNORDERED_ACCESS_VIEWS)
{
WARN("Disabling ARB_shader_atomic_counters.\n");
gl_info->supported[ARB_SHADER_ATOMIC_COUNTERS] = FALSE;
}
}
if (gl_info->supported[ARB_TRANSFORM_FEEDBACK3])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_VERTEX_STREAMS, &gl_max);
TRACE("Max vertex streams: %d.\n", gl_max);
}
if (gl_info->supported[NV_LIGHT_MAX_EXPONENT])
gl_info->gl_ops.gl.p_glGetFloatv(GL_MAX_SHININESS_NV, &gl_info->limits.shininess);
else
gl_info->limits.shininess = 128.0f;
if (gl_info->supported[ARB_FRAMEBUFFER_OBJECT] || gl_info->supported[EXT_FRAMEBUFFER_MULTISAMPLE])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_SAMPLES, &gl_max);
gl_info->limits.samples = gl_max;
}
if (gl_info->supported[ARB_FRAMEBUFFER_NO_ATTACHMENTS])
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAMEBUFFER_WIDTH, &gl_max);
gl_info->limits.framebuffer_width = gl_max;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_MAX_FRAMEBUFFER_HEIGHT, &gl_max);
gl_info->limits.framebuffer_height = gl_max;
}
else
{
gl_info->limits.framebuffer_width = gl_info->limits.texture_size;
gl_info->limits.framebuffer_height = gl_info->limits.texture_size;
}
gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL] =
min(gl_info->limits.samplers[WINED3D_SHADER_TYPE_PIXEL], MAX_GL_FRAGMENT_SAMPLERS);
sampler_count = 0;
for (i = 0; i < WINED3D_SHADER_TYPE_GRAPHICS_COUNT; ++i)
sampler_count += gl_info->limits.samplers[i];
if (gl_info->supported[WINED3D_GL_VERSION_3_2] && gl_info->limits.combined_samplers < sampler_count)
{
/* The minimum value for GL_MAX_COMBINED_TEXTURE_IMAGE_UNITS in OpenGL
* 3.2 is 48 (16 per stage). When tessellation shaders are supported
* the minimum value is increased to 80. */
WARN("Graphics pipeline sampler count %u is greater than combined sampler count %u.\n",
sampler_count, gl_info->limits.combined_samplers);
for (i = 0; i < WINED3D_SHADER_TYPE_GRAPHICS_COUNT; ++i)
gl_info->limits.samplers[i] = min(gl_info->limits.samplers[i], 16);
}
/* A majority of OpenGL implementations allow us to statically partition
* the set of texture bindings into six separate sets. */
gl_info->limits.graphics_samplers = gl_info->limits.combined_samplers;
sampler_count = 0;
for (i = 0; i < WINED3D_SHADER_TYPE_COUNT; ++i)
sampler_count += gl_info->limits.samplers[i];
if (gl_info->limits.combined_samplers >= sampler_count)
gl_info->limits.graphics_samplers -= gl_info->limits.samplers[WINED3D_SHADER_TYPE_COMPUTE];
}
/* Context activation is done by the caller. */
static BOOL wined3d_adapter_init_gl_caps(struct wined3d_adapter *adapter,
struct wined3d_caps_gl_ctx *caps_gl_ctx, unsigned int wined3d_creation_flags)
{
static const struct
{
enum wined3d_gl_extension extension;
DWORD min_gl_version;
}
core_extensions[] =
{
{EXT_TEXTURE3D, MAKEDWORD_VERSION(1, 2)},
{ARB_MULTISAMPLE, MAKEDWORD_VERSION(1, 3)},
{ARB_MULTITEXTURE, MAKEDWORD_VERSION(1, 3)},
{ARB_TEXTURE_BORDER_CLAMP, MAKEDWORD_VERSION(1, 3)},
{ARB_TEXTURE_COMPRESSION, MAKEDWORD_VERSION(1, 3)},
{ARB_TEXTURE_CUBE_MAP, MAKEDWORD_VERSION(1, 3)},
{ARB_DEPTH_TEXTURE, MAKEDWORD_VERSION(1, 4)},
{ARB_POINT_PARAMETERS, MAKEDWORD_VERSION(1, 4)},
{ARB_SHADOW, MAKEDWORD_VERSION(1, 4)},
{ARB_TEXTURE_MIRRORED_REPEAT, MAKEDWORD_VERSION(1, 4)},
{EXT_BLEND_COLOR, MAKEDWORD_VERSION(1, 4)},
{EXT_BLEND_FUNC_SEPARATE, MAKEDWORD_VERSION(1, 4)},
{EXT_BLEND_MINMAX, MAKEDWORD_VERSION(1, 4)},
{EXT_BLEND_SUBTRACT, MAKEDWORD_VERSION(1, 4)},
{EXT_STENCIL_WRAP, MAKEDWORD_VERSION(1, 4)},
{NV_POINT_SPRITE, MAKEDWORD_VERSION(1, 4)},
{ARB_OCCLUSION_QUERY, MAKEDWORD_VERSION(1, 5)},
{ARB_VERTEX_BUFFER_OBJECT, MAKEDWORD_VERSION(1, 5)},
{ARB_DRAW_BUFFERS, MAKEDWORD_VERSION(2, 0)},
{ARB_FRAGMENT_SHADER, MAKEDWORD_VERSION(2, 0)},
{ARB_SHADING_LANGUAGE_100, MAKEDWORD_VERSION(2, 0)},
{ARB_TEXTURE_NON_POWER_OF_TWO, MAKEDWORD_VERSION(2, 0)},
{ARB_VERTEX_SHADER, MAKEDWORD_VERSION(2, 0)},
{EXT_BLEND_EQUATION_SEPARATE, MAKEDWORD_VERSION(2, 0)},
{ARB_PIXEL_BUFFER_OBJECT, MAKEDWORD_VERSION(2, 1)},
{EXT_TEXTURE_SRGB, MAKEDWORD_VERSION(2, 1)},
{ARB_COLOR_BUFFER_FLOAT, MAKEDWORD_VERSION(3, 0)},
{ARB_DEPTH_BUFFER_FLOAT, MAKEDWORD_VERSION(3, 0)},
{ARB_FRAMEBUFFER_OBJECT, MAKEDWORD_VERSION(3, 0)},
{ARB_FRAMEBUFFER_SRGB, MAKEDWORD_VERSION(3, 0)},
{ARB_HALF_FLOAT_PIXEL, MAKEDWORD_VERSION(3, 0)},
{ARB_HALF_FLOAT_VERTEX, MAKEDWORD_VERSION(3, 0)},
{ARB_MAP_BUFFER_RANGE, MAKEDWORD_VERSION(3, 0)},
{ARB_TEXTURE_COMPRESSION_RGTC, MAKEDWORD_VERSION(3, 0)},
{ARB_TEXTURE_FLOAT, MAKEDWORD_VERSION(3, 0)},
{ARB_TEXTURE_RG, MAKEDWORD_VERSION(3, 0)},
{EXT_DRAW_BUFFERS2, MAKEDWORD_VERSION(3, 0)},
{EXT_PACKED_FLOAT, MAKEDWORD_VERSION(3, 0)},
{EXT_TEXTURE_ARRAY, MAKEDWORD_VERSION(3, 0)},
{EXT_TEXTURE_INTEGER, MAKEDWORD_VERSION(3, 0)},
{EXT_TEXTURE_SHARED_EXPONENT, MAKEDWORD_VERSION(3, 0)},
/* We don't want to enable EXT_GPU_SHADER4: even though similar
* functionality is available in core GL 3.0 / GLSL 1.30, it's different
* enough that reusing the same flag for the new features hurts more
* than it helps. */
/* EXT_framebuffer_object, EXT_framebuffer_blit,
* EXT_framebuffer_multisample and EXT_packed_depth_stencil
* are integrated into ARB_framebuffer_object. */
{ARB_COPY_BUFFER, MAKEDWORD_VERSION(3, 1)},
{ARB_DRAW_INSTANCED, MAKEDWORD_VERSION(3, 1)},
{ARB_TEXTURE_BUFFER_OBJECT, MAKEDWORD_VERSION(3, 1)},
{ARB_UNIFORM_BUFFER_OBJECT, MAKEDWORD_VERSION(3, 1)},
{EXT_TEXTURE_SNORM, MAKEDWORD_VERSION(3, 1)},
/* We don't need or want GL_ARB_texture_rectangle (core in 3.1). */
{ARB_DEPTH_CLAMP, MAKEDWORD_VERSION(3, 2)},
{ARB_DRAW_ELEMENTS_BASE_VERTEX, MAKEDWORD_VERSION(3, 2)},
/* ARB_geometry_shader4 exposes a somewhat different API compared to 3.2
* core geometry shaders so it's not really correct to expose the
* extension for core-only support. */
{ARB_FRAGMENT_COORD_CONVENTIONS, MAKEDWORD_VERSION(3, 2)},
{ARB_PROVOKING_VERTEX, MAKEDWORD_VERSION(3, 2)},
{ARB_SEAMLESS_CUBE_MAP, MAKEDWORD_VERSION(3, 2)},
{ARB_SYNC, MAKEDWORD_VERSION(3, 2)},
{ARB_TEXTURE_MULTISAMPLE, MAKEDWORD_VERSION(3, 2)},
{ARB_VERTEX_ARRAY_BGRA, MAKEDWORD_VERSION(3, 2)},
{ARB_BLEND_FUNC_EXTENDED, MAKEDWORD_VERSION(3, 3)},
{ARB_EXPLICIT_ATTRIB_LOCATION, MAKEDWORD_VERSION(3, 3)},
{ARB_INSTANCED_ARRAYS, MAKEDWORD_VERSION(3, 3)},
{ARB_SAMPLER_OBJECTS, MAKEDWORD_VERSION(3, 3)},
{ARB_SHADER_BIT_ENCODING, MAKEDWORD_VERSION(3, 3)},
{ARB_TEXTURE_RGB10_A2UI, MAKEDWORD_VERSION(3, 3)},
{ARB_TEXTURE_SWIZZLE, MAKEDWORD_VERSION(3, 3)},
{ARB_TIMER_QUERY, MAKEDWORD_VERSION(3, 3)},
{ARB_VERTEX_TYPE_2_10_10_10_REV, MAKEDWORD_VERSION(3, 3)},
{ARB_DRAW_INDIRECT, MAKEDWORD_VERSION(4, 0)},
{ARB_GPU_SHADER5, MAKEDWORD_VERSION(4, 0)},
{ARB_SAMPLE_SHADING, MAKEDWORD_VERSION(4, 0)},
{ARB_TESSELLATION_SHADER, MAKEDWORD_VERSION(4, 0)},
{ARB_TEXTURE_CUBE_MAP_ARRAY, MAKEDWORD_VERSION(4, 0)},
{ARB_TEXTURE_GATHER, MAKEDWORD_VERSION(4, 0)},
{ARB_TRANSFORM_FEEDBACK2, MAKEDWORD_VERSION(4, 0)},
{ARB_TRANSFORM_FEEDBACK3, MAKEDWORD_VERSION(4, 0)},
{ARB_ES2_COMPATIBILITY, MAKEDWORD_VERSION(4, 1)},
{ARB_VIEWPORT_ARRAY, MAKEDWORD_VERSION(4, 1)},
{ARB_BASE_INSTANCE, MAKEDWORD_VERSION(4, 2)},
{ARB_CONSERVATIVE_DEPTH, MAKEDWORD_VERSION(4, 2)},
{ARB_INTERNALFORMAT_QUERY, MAKEDWORD_VERSION(4, 2)},
{ARB_MAP_BUFFER_ALIGNMENT, MAKEDWORD_VERSION(4, 2)},
{ARB_SHADER_ATOMIC_COUNTERS, MAKEDWORD_VERSION(4, 2)},
{ARB_SHADER_IMAGE_LOAD_STORE, MAKEDWORD_VERSION(4, 2)},
{ARB_SHADING_LANGUAGE_420PACK, MAKEDWORD_VERSION(4, 2)},
{ARB_SHADING_LANGUAGE_PACKING, MAKEDWORD_VERSION(4, 2)},
{ARB_TEXTURE_COMPRESSION_BPTC, MAKEDWORD_VERSION(4, 2)},
{ARB_TEXTURE_STORAGE, MAKEDWORD_VERSION(4, 2)},
{ARB_CLEAR_BUFFER_OBJECT, MAKEDWORD_VERSION(4, 3)},
{ARB_COMPUTE_SHADER, MAKEDWORD_VERSION(4, 3)},
{ARB_COPY_IMAGE, MAKEDWORD_VERSION(4, 3)},
{ARB_DEBUG_OUTPUT, MAKEDWORD_VERSION(4, 3)},
{ARB_ES3_COMPATIBILITY, MAKEDWORD_VERSION(4, 3)},
{ARB_FRAGMENT_LAYER_VIEWPORT, MAKEDWORD_VERSION(4, 3)},
{ARB_FRAMEBUFFER_NO_ATTACHMENTS, MAKEDWORD_VERSION(4, 3)},
{ARB_INTERNALFORMAT_QUERY2, MAKEDWORD_VERSION(4, 3)},
{ARB_SHADER_IMAGE_SIZE, MAKEDWORD_VERSION(4, 3)},
{ARB_SHADER_STORAGE_BUFFER_OBJECT, MAKEDWORD_VERSION(4, 3)},
{ARB_STENCIL_TEXTURING, MAKEDWORD_VERSION(4, 3)},
{ARB_TEXTURE_BUFFER_RANGE, MAKEDWORD_VERSION(4, 3)},
{ARB_TEXTURE_QUERY_LEVELS, MAKEDWORD_VERSION(4, 3)},
{ARB_TEXTURE_STORAGE_MULTISAMPLE, MAKEDWORD_VERSION(4, 2)},
{ARB_TEXTURE_VIEW, MAKEDWORD_VERSION(4, 3)},
{ARB_BUFFER_STORAGE, MAKEDWORD_VERSION(4, 4)},
{ARB_CLEAR_TEXTURE, MAKEDWORD_VERSION(4, 4)},
{ARB_QUERY_BUFFER_OBJECT, MAKEDWORD_VERSION(4, 4)},
{ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE, MAKEDWORD_VERSION(4, 4)},
{ARB_CLIP_CONTROL, MAKEDWORD_VERSION(4, 5)},
{ARB_CULL_DISTANCE, MAKEDWORD_VERSION(4, 5)},
{ARB_DERIVATIVE_CONTROL, MAKEDWORD_VERSION(4, 5)},
{ARB_SHADER_TEXTURE_IMAGE_SAMPLES, MAKEDWORD_VERSION(4, 5)},
{ARB_PIPELINE_STATISTICS_QUERY, MAKEDWORD_VERSION(4, 6)},
{ARB_POLYGON_OFFSET_CLAMP, MAKEDWORD_VERSION(4, 6)},
{ARB_TEXTURE_FILTER_ANISOTROPIC, MAKEDWORD_VERSION(4, 6)},
};
struct wined3d_driver_info *driver_info = &adapter->driver_info;
const char *gl_vendor_str, *gl_renderer_str, *gl_version_str;
const struct wined3d_gpu_description *gpu_description;
struct wined3d_gl_info *gl_info = &adapter->gl_info;
const char *WGL_Extensions = NULL;
enum wined3d_gl_vendor gl_vendor;
DWORD gl_version, gl_ext_emul_mask;
GLint context_profile = 0;
UINT64 vram_bytes = 0;
unsigned int i, j;
HDC hdc;
TRACE("adapter %p.\n", adapter);
gl_renderer_str = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_RENDERER);
TRACE("GL_RENDERER: %s.\n", debugstr_a(gl_renderer_str));
if (!gl_renderer_str)
{
ERR("Received a NULL GL_RENDERER.\n");
return FALSE;
}
gl_vendor_str = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_VENDOR);
TRACE("GL_VENDOR: %s.\n", debugstr_a(gl_vendor_str));
if (!gl_vendor_str)
{
ERR("Received a NULL GL_VENDOR.\n");
return FALSE;
}
/* Parse the GL_VERSION field into major and minor information */
gl_version_str = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_VERSION);
TRACE("GL_VERSION: %s.\n", debugstr_a(gl_version_str));
if (!gl_version_str)
{
ERR("Received a NULL GL_VERSION.\n");
return FALSE;
}
gl_version = wined3d_parse_gl_version(gl_version_str);
load_gl_funcs(gl_info);
memset(gl_info->supported, 0, sizeof(gl_info->supported));
gl_info->supported[WINED3D_GL_EXT_NONE] = TRUE;
if (gl_version >= MAKEDWORD_VERSION(3, 2))
{
gl_info->gl_ops.gl.p_glGetIntegerv(GL_CONTEXT_PROFILE_MASK, &context_profile);
checkGLcall("Querying context profile");
}
if (context_profile & GL_CONTEXT_CORE_PROFILE_BIT)
TRACE("Got a core profile context.\n");
else
gl_info->supported[WINED3D_GL_LEGACY_CONTEXT] = TRUE;
TRACE("GL extensions reported:\n");
if (gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
{
const char *gl_extensions = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_EXTENSIONS);
if (!gl_extensions)
{
ERR("Received a NULL GL_EXTENSIONS.\n");
return FALSE;
}
parse_extension_string(gl_info, gl_extensions, gl_extension_map, ARRAY_SIZE(gl_extension_map));
}
else
{
enumerate_gl_extensions(gl_info, gl_extension_map, ARRAY_SIZE(gl_extension_map));
}
hdc = wglGetCurrentDC();
/* Not all GL drivers might offer WGL extensions e.g. VirtualBox. */
if (GL_EXTCALL(wglGetExtensionsStringARB))
WGL_Extensions = (const char *)GL_EXTCALL(wglGetExtensionsStringARB(hdc));
if (!WGL_Extensions)
WARN("WGL extensions not supported.\n");
else
parse_extension_string(gl_info, WGL_Extensions, wgl_extension_map, ARRAY_SIZE(wgl_extension_map));
for (i = 0; i < ARRAY_SIZE(core_extensions); ++i)
{
if (!gl_info->supported[core_extensions[i].extension]
&& gl_version >= core_extensions[i].min_gl_version)
{
for (j = 0; j < ARRAY_SIZE(gl_extension_map); ++j)
if (gl_extension_map[j].extension == core_extensions[i].extension)
break;
if (j < ARRAY_SIZE(gl_extension_map))
{
TRACE("GL CORE: %s support.\n", gl_extension_map[j].extension_string);
gl_info->supported[core_extensions[i].extension] = TRUE;
}
else
{
FIXME("GL extension %u not in the GL extensions map.\n", core_extensions[i].extension);
}
}
}
if (gl_info->supported[EXT_BLEND_MINMAX] || gl_info->supported[EXT_BLEND_SUBTRACT])
gl_info->supported[WINED3D_GL_BLEND_EQUATION] = TRUE;
if (gl_version >= MAKEDWORD_VERSION(2, 0))
{
gl_info->supported[WINED3D_GL_VERSION_2_0] = TRUE;
/* We want to use the core APIs for two-sided stencil in GL 2.0. */
gl_info->supported[EXT_STENCIL_TWO_SIDE] = FALSE;
}
if (gl_version >= MAKEDWORD_VERSION(3, 2))
gl_info->supported[WINED3D_GL_VERSION_3_2] = TRUE;
/* All the points are actually point sprites in core contexts, the APIs from
* ARB_point_sprite are not supported anymore. */
if (!gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
gl_info->supported[ARB_POINT_SPRITE] = FALSE;
if (gl_info->supported[APPLE_FENCE])
{
/* GL_NV_fence and GL_APPLE_fence provide the same functionality basically.
* The apple extension interacts with some other apple exts. Disable the NV
* extension if the apple one is support to prevent confusion in other parts
* of the code. */
gl_info->supported[NV_FENCE] = FALSE;
}
if (gl_info->supported[APPLE_FLOAT_PIXELS])
{
/* GL_APPLE_float_pixels == GL_ARB_texture_float + GL_ARB_half_float_pixel
*
* The enums are the same:
* GL_RGBA16F_ARB = GL_RGBA_FLOAT16_APPLE = 0x881a
* GL_RGB16F_ARB = GL_RGB_FLOAT16_APPLE = 0x881b
* GL_RGBA32F_ARB = GL_RGBA_FLOAT32_APPLE = 0x8814
* GL_RGB32F_ARB = GL_RGB_FLOAT32_APPLE = 0x8815
* GL_HALF_FLOAT_ARB = GL_HALF_APPLE = 0x140b
*/
if (!gl_info->supported[ARB_TEXTURE_FLOAT])
{
TRACE(" IMPLIED: GL_ARB_texture_float support (by GL_APPLE_float_pixels).\n");
gl_info->supported[ARB_TEXTURE_FLOAT] = TRUE;
}
if (!gl_info->supported[ARB_HALF_FLOAT_PIXEL])
{
TRACE(" IMPLIED: GL_ARB_half_float_pixel support (by GL_APPLE_float_pixels).\n");
gl_info->supported[ARB_HALF_FLOAT_PIXEL] = TRUE;
}
}
if (gl_info->supported[ARB_MAP_BUFFER_RANGE])
{
/* GL_ARB_map_buffer_range and GL_APPLE_flush_buffer_range provide the same
* functionality. Prefer the ARB extension */
gl_info->supported[APPLE_FLUSH_BUFFER_RANGE] = FALSE;
}
if (gl_info->supported[ARB_TEXTURE_CUBE_MAP])
{
TRACE(" IMPLIED: NVIDIA (NV) Texture Gen Reflection support.\n");
gl_info->supported[NV_TEXGEN_REFLECTION] = TRUE;
}
if (!gl_info->supported[EXT_TEXTURE_COMPRESSION_RGTC] && gl_info->supported[ARB_TEXTURE_COMPRESSION_RGTC])
{
TRACE(" IMPLIED: EXT_texture_compression_rgtc support (by ARB_texture_compression_rgtc).\n");
gl_info->supported[EXT_TEXTURE_COMPRESSION_RGTC] = TRUE;
}
if (!gl_info->supported[ARB_TEXTURE_COMPRESSION_RGTC] && gl_info->supported[EXT_TEXTURE_COMPRESSION_RGTC])
{
TRACE(" IMPLIED: ARB_texture_compression_rgtc support (by EXT_texture_compression_rgtc).\n");
gl_info->supported[ARB_TEXTURE_COMPRESSION_RGTC] = TRUE;
}
if (gl_info->supported[ARB_TEXTURE_COMPRESSION_RGTC] && !gl_info->supported[ARB_TEXTURE_RG])
{
TRACE("ARB_texture_rg not supported, disabling ARB_texture_compression_rgtc.\n");
gl_info->supported[ARB_TEXTURE_COMPRESSION_RGTC] = FALSE;
}
if (gl_info->supported[NV_TEXTURE_SHADER2])
{
if (gl_info->supported[NV_REGISTER_COMBINERS])
{
/* Also disable ATI_FRAGMENT_SHADER if register combiners and texture_shader2
* are supported. The nv extensions provide the same functionality as the
* ATI one, and a bit more(signed pixelformats). */
gl_info->supported[ATI_FRAGMENT_SHADER] = FALSE;
}
}
if (gl_info->supported[ARB_TEXTURE_NON_POWER_OF_TWO])
{
/* If we have full NP2 texture support, disable
* GL_ARB_texture_rectangle because we will never use it.
* This saves a few redundant glDisable calls. */
gl_info->supported[ARB_TEXTURE_RECTANGLE] = FALSE;
}
if (gl_info->supported[ATI_FRAGMENT_SHADER])
{
/* Disable NV_register_combiners and fragment shader if this is supported.
* generally the NV extensions are preferred over the ATI ones, and this
* extension is disabled if register_combiners and texture_shader2 are both
* supported. So we reach this place only if we have incomplete NV dxlevel 8
* fragment processing support. */
gl_info->supported[NV_REGISTER_COMBINERS] = FALSE;
gl_info->supported[NV_REGISTER_COMBINERS2] = FALSE;
gl_info->supported[NV_TEXTURE_SHADER] = FALSE;
gl_info->supported[NV_TEXTURE_SHADER2] = FALSE;
}
if (gl_info->supported[NV_HALF_FLOAT])
{
/* GL_ARB_half_float_vertex is a subset of GL_NV_half_float. */
gl_info->supported[ARB_HALF_FLOAT_VERTEX] = TRUE;
}
if (gl_info->supported[ARB_FRAMEBUFFER_SRGB] && !gl_info->supported[EXT_TEXTURE_SRGB_DECODE])
{
/* Current wined3d sRGB infrastructure requires EXT_texture_sRGB_decode
* for GL_ARB_framebuffer_sRGB support (without EXT_texture_sRGB_decode
* we never render to sRGB surfaces). */
TRACE("EXT_texture_sRGB_decode is not supported, disabling ARB_framebuffer_sRGB.\n");
gl_info->supported[ARB_FRAMEBUFFER_SRGB] = FALSE;
}
if (gl_info->supported[ARB_OCCLUSION_QUERY])
{
GLint counter_bits;
GL_EXTCALL(glGetQueryiv(GL_SAMPLES_PASSED, GL_QUERY_COUNTER_BITS, &counter_bits));
TRACE("Occlusion query counter has %d bits.\n", counter_bits);
if (!counter_bits)
gl_info->supported[ARB_OCCLUSION_QUERY] = FALSE;
}
if (gl_info->supported[ARB_TIMER_QUERY])
{
GLint counter_bits;
GL_EXTCALL(glGetQueryiv(GL_TIMESTAMP, GL_QUERY_COUNTER_BITS, &counter_bits));
TRACE("Timestamp query counter has %d bits.\n", counter_bits);
if (!counter_bits)
gl_info->supported[ARB_TIMER_QUERY] = FALSE;
}
if (gl_version >= MAKEDWORD_VERSION(3, 0))
{
GLint counter_bits;
gl_info->supported[WINED3D_GL_PRIMITIVE_QUERY] = TRUE;
GL_EXTCALL(glGetQueryiv(GL_PRIMITIVES_GENERATED, GL_QUERY_COUNTER_BITS, &counter_bits));
TRACE("Primitives query counter has %d bits.\n", counter_bits);
if (!counter_bits)
gl_info->supported[WINED3D_GL_PRIMITIVE_QUERY] = FALSE;
GL_EXTCALL(glGetQueryiv(GL_TRANSFORM_FEEDBACK_PRIMITIVES_WRITTEN, GL_QUERY_COUNTER_BITS, &counter_bits));
TRACE("Transform feedback primitives query counter has %d bits.\n", counter_bits);
if (!counter_bits)
gl_info->supported[WINED3D_GL_PRIMITIVE_QUERY] = FALSE;
}
if (gl_info->supported[ARB_VIEWPORT_ARRAY])
{
GLint subpixel_bits;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_VIEWPORT_SUBPIXEL_BITS, &subpixel_bits);
TRACE("Viewport supports %d subpixel bits.\n", subpixel_bits);
if (subpixel_bits < 8 && gl_info->supported[ARB_CLIP_CONTROL])
{
TRACE("Disabling ARB_clip_control because viewport subpixel bits < 8.\n");
gl_info->supported[ARB_CLIP_CONTROL] = FALSE;
}
}
if (gl_info->supported[ARB_CLIP_CONTROL] && !gl_info->supported[ARB_VIEWPORT_ARRAY])
{
/* When using ARB_clip_control we need the float viewport parameters
* introduced by ARB_viewport_array to take care of the shifted pixel
* coordinates. */
TRACE("Disabling ARB_clip_control because ARB_viewport_array is not supported.\n");
gl_info->supported[ARB_CLIP_CONTROL] = FALSE;
}
if (gl_info->supported[ARB_STENCIL_TEXTURING] && !gl_info->supported[ARB_TEXTURE_SWIZZLE])
{
/* The stencil value needs to be placed in the green channel. */
TRACE("Disabling ARB_stencil_texturing because ARB_texture_swizzle is not supported.\n");
gl_info->supported[ARB_STENCIL_TEXTURING] = FALSE;
}
if (!gl_info->supported[ATI_TEXTURE_MIRROR_ONCE] && gl_info->supported[EXT_TEXTURE_MIRROR_CLAMP])
{
TRACE(" IMPLIED: ATI_texture_mirror_once support (by EXT_texture_mirror_clamp).\n");
gl_info->supported[ATI_TEXTURE_MIRROR_ONCE] = TRUE;
}
if (!gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE] && gl_info->supported[ATI_TEXTURE_MIRROR_ONCE])
{
TRACE(" IMPLIED: ARB_texture_mirror_clamp_to_edge support (by ATI_texture_mirror_once).\n");
gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE] = TRUE;
}
if (gl_info->supported[ARB_TEXTURE_STORAGE] && gl_info->supported[APPLE_YCBCR_422])
{
/* AFAIK APPLE_ycbcr_422 is only available in legacy contexts so we shouldn't ever hit this. */
ERR("Disabling APPLE_ycbcr_422 because of ARB_texture_storage.\n");
gl_info->supported[APPLE_YCBCR_422] = FALSE;
}
if (gl_info->supported[ARB_DRAW_INDIRECT] && !gl_info->supported[ARB_BASE_INSTANCE])
{
/* If ARB_base_instance is not supported the baseInstance field
* in indirect draw parameters must be 0 or behavior is undefined.
*/
WARN("Disabling ARB_draw_indirect because ARB_base_instance is not supported.\n");
gl_info->supported[ARB_DRAW_INDIRECT] = FALSE;
}
if (gl_info->supported[ARB_TEXTURE_MULTISAMPLE] && !wined3d_settings.multisample_textures)
gl_info->supported[ARB_TEXTURE_MULTISAMPLE] = FALSE;
if (gl_info->supported[ARB_TEXTURE_MULTISAMPLE] && !gl_info->supported[ARB_TEXTURE_STORAGE_MULTISAMPLE])
{
WARN("Disabling ARB_texture_multisample because immutable storage is not supported.\n");
gl_info->supported[ARB_TEXTURE_MULTISAMPLE] = FALSE;
}
wined3d_adapter_init_limits(gl_info);
if (gl_info->supported[ARB_VERTEX_PROGRAM] && test_arb_vs_offset_limit(gl_info))
gl_info->quirks |= WINED3D_QUIRK_ARB_VS_OFFSET_LIMIT;
if (gl_info->supported[ARB_SHADING_LANGUAGE_100])
{
const char *str = (const char *)gl_info->gl_ops.gl.p_glGetString(GL_SHADING_LANGUAGE_VERSION_ARB);
unsigned int major, minor;
TRACE("GLSL version string: %s.\n", debugstr_a(str));
/* The format of the GLSL version string is "major.minor[.release] [vendor info]". */
sscanf(str, "%u.%u", &major, &minor);
gl_info->glsl_version = MAKEDWORD_VERSION(major, minor);
if (gl_info->glsl_version >= MAKEDWORD_VERSION(1, 30))
gl_info->supported[WINED3D_GLSL_130] = TRUE;
}
checkGLcall("extension detection");
adapter->shader_backend = select_shader_backend(gl_info);
adapter->vertex_pipe = select_vertex_implementation(gl_info, adapter->shader_backend);
adapter->fragment_pipe = select_fragment_implementation(gl_info, adapter->shader_backend);
if (gl_info->supported[ARB_FRAMEBUFFER_OBJECT])
{
gl_info->fbo_ops.glIsRenderbuffer = gl_info->gl_ops.ext.p_glIsRenderbuffer;
gl_info->fbo_ops.glBindRenderbuffer = gl_info->gl_ops.ext.p_glBindRenderbuffer;
gl_info->fbo_ops.glDeleteRenderbuffers = gl_info->gl_ops.ext.p_glDeleteRenderbuffers;
gl_info->fbo_ops.glGenRenderbuffers = gl_info->gl_ops.ext.p_glGenRenderbuffers;
gl_info->fbo_ops.glRenderbufferStorage = gl_info->gl_ops.ext.p_glRenderbufferStorage;
gl_info->fbo_ops.glRenderbufferStorageMultisample = gl_info->gl_ops.ext.p_glRenderbufferStorageMultisample;
gl_info->fbo_ops.glGetRenderbufferParameteriv = gl_info->gl_ops.ext.p_glGetRenderbufferParameteriv;
gl_info->fbo_ops.glIsFramebuffer = gl_info->gl_ops.ext.p_glIsFramebuffer;
gl_info->fbo_ops.glBindFramebuffer = gl_info->gl_ops.ext.p_glBindFramebuffer;
gl_info->fbo_ops.glDeleteFramebuffers = gl_info->gl_ops.ext.p_glDeleteFramebuffers;
gl_info->fbo_ops.glGenFramebuffers = gl_info->gl_ops.ext.p_glGenFramebuffers;
gl_info->fbo_ops.glCheckFramebufferStatus = gl_info->gl_ops.ext.p_glCheckFramebufferStatus;
gl_info->fbo_ops.glFramebufferTexture1D = gl_info->gl_ops.ext.p_glFramebufferTexture1D;
gl_info->fbo_ops.glFramebufferTexture2D = gl_info->gl_ops.ext.p_glFramebufferTexture2D;
gl_info->fbo_ops.glFramebufferTexture3D = gl_info->gl_ops.ext.p_glFramebufferTexture3D;
gl_info->fbo_ops.glFramebufferTextureLayer = gl_info->gl_ops.ext.p_glFramebufferTextureLayer;
gl_info->fbo_ops.glFramebufferRenderbuffer = gl_info->gl_ops.ext.p_glFramebufferRenderbuffer;
gl_info->fbo_ops.glGetFramebufferAttachmentParameteriv
= gl_info->gl_ops.ext.p_glGetFramebufferAttachmentParameteriv;
gl_info->fbo_ops.glBlitFramebuffer = gl_info->gl_ops.ext.p_glBlitFramebuffer;
gl_info->fbo_ops.glGenerateMipmap = gl_info->gl_ops.ext.p_glGenerateMipmap;
gl_info->fbo_ops.glFramebufferTexture = gl_info->gl_ops.ext.p_glFramebufferTexture;
}
else
{
if (gl_info->supported[EXT_FRAMEBUFFER_OBJECT])
{
gl_info->fbo_ops.glIsRenderbuffer = gl_info->gl_ops.ext.p_glIsRenderbufferEXT;
gl_info->fbo_ops.glBindRenderbuffer = gl_info->gl_ops.ext.p_glBindRenderbufferEXT;
gl_info->fbo_ops.glDeleteRenderbuffers = gl_info->gl_ops.ext.p_glDeleteRenderbuffersEXT;
gl_info->fbo_ops.glGenRenderbuffers = gl_info->gl_ops.ext.p_glGenRenderbuffersEXT;
gl_info->fbo_ops.glRenderbufferStorage = gl_info->gl_ops.ext.p_glRenderbufferStorageEXT;
gl_info->fbo_ops.glGetRenderbufferParameteriv = gl_info->gl_ops.ext.p_glGetRenderbufferParameterivEXT;
gl_info->fbo_ops.glIsFramebuffer = gl_info->gl_ops.ext.p_glIsFramebufferEXT;
gl_info->fbo_ops.glBindFramebuffer = gl_info->gl_ops.ext.p_glBindFramebufferEXT;
gl_info->fbo_ops.glDeleteFramebuffers = gl_info->gl_ops.ext.p_glDeleteFramebuffersEXT;
gl_info->fbo_ops.glGenFramebuffers = gl_info->gl_ops.ext.p_glGenFramebuffersEXT;
gl_info->fbo_ops.glCheckFramebufferStatus = gl_info->gl_ops.ext.p_glCheckFramebufferStatusEXT;
gl_info->fbo_ops.glFramebufferTexture1D = gl_info->gl_ops.ext.p_glFramebufferTexture1DEXT;
gl_info->fbo_ops.glFramebufferTexture2D = gl_info->gl_ops.ext.p_glFramebufferTexture2DEXT;
gl_info->fbo_ops.glFramebufferTexture3D = gl_info->gl_ops.ext.p_glFramebufferTexture3DEXT;
gl_info->fbo_ops.glFramebufferRenderbuffer = gl_info->gl_ops.ext.p_glFramebufferRenderbufferEXT;
gl_info->fbo_ops.glGetFramebufferAttachmentParameteriv
= gl_info->gl_ops.ext.p_glGetFramebufferAttachmentParameterivEXT;
gl_info->fbo_ops.glGenerateMipmap = gl_info->gl_ops.ext.p_glGenerateMipmapEXT;
}
else if (wined3d_settings.offscreen_rendering_mode == ORM_FBO)
{
WARN_(d3d_perf)("Framebuffer objects not supported, falling back to backbuffer offscreen rendering mode.\n");
wined3d_settings.offscreen_rendering_mode = ORM_BACKBUFFER;
}
if (gl_info->supported[ARB_GEOMETRY_SHADER4])
{
gl_info->fbo_ops.glFramebufferTexture = gl_info->gl_ops.ext.p_glFramebufferTextureARB;
gl_info->fbo_ops.glFramebufferTextureLayer = gl_info->gl_ops.ext.p_glFramebufferTextureLayerARB;
}
if (gl_info->supported[EXT_FRAMEBUFFER_BLIT])
{
gl_info->fbo_ops.glBlitFramebuffer = gl_info->gl_ops.ext.p_glBlitFramebufferEXT;
}
if (gl_info->supported[EXT_FRAMEBUFFER_MULTISAMPLE])
{
gl_info->fbo_ops.glRenderbufferStorageMultisample
= gl_info->gl_ops.ext.p_glRenderbufferStorageMultisampleEXT;
}
}
gl_info->wrap_lookup[WINED3D_TADDRESS_WRAP - WINED3D_TADDRESS_WRAP] = GL_REPEAT;
gl_info->wrap_lookup[WINED3D_TADDRESS_MIRROR - WINED3D_TADDRESS_WRAP] =
gl_info->supported[ARB_TEXTURE_MIRRORED_REPEAT] ? GL_MIRRORED_REPEAT_ARB : GL_REPEAT;
gl_info->wrap_lookup[WINED3D_TADDRESS_CLAMP - WINED3D_TADDRESS_WRAP] = GL_CLAMP_TO_EDGE;
gl_info->wrap_lookup[WINED3D_TADDRESS_BORDER - WINED3D_TADDRESS_WRAP] =
gl_info->supported[ARB_TEXTURE_BORDER_CLAMP] ? GL_CLAMP_TO_BORDER_ARB : GL_REPEAT;
gl_info->wrap_lookup[WINED3D_TADDRESS_MIRROR_ONCE - WINED3D_TADDRESS_WRAP] =
gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE] ? GL_MIRROR_CLAMP_TO_EDGE : GL_REPEAT;
if (!gl_info->supported[WINED3D_GL_LEGACY_CONTEXT])
{
GLuint vao;
GL_EXTCALL(glGenVertexArrays(1, &vao));
GL_EXTCALL(glBindVertexArray(vao));
checkGLcall("creating VAO");
}
gl_vendor = wined3d_guess_gl_vendor(gl_info, gl_vendor_str, gl_renderer_str, gl_version_str);
TRACE("Guessed GL vendor %#x.\n", gl_vendor);
if (!(gpu_description = query_gpu_description(gl_info, &vram_bytes)))
{
enum wined3d_feature_level feature_level;
struct fragment_caps fragment_caps;
enum wined3d_pci_vendor vendor;
enum wined3d_pci_device device;
struct shader_caps shader_caps;
adapter->shader_backend->shader_get_caps(adapter, &shader_caps);
adapter->fragment_pipe->get_caps(adapter, &fragment_caps);
feature_level = feature_level_from_caps(gl_info, &shader_caps, &fragment_caps);
vendor = wined3d_guess_card_vendor(gl_vendor_str, gl_renderer_str);
TRACE("Guessed vendor PCI ID 0x%04x.\n", vendor);
device = wined3d_guess_card(feature_level, gl_renderer_str, &gl_vendor, &vendor);
TRACE("Guessed device PCI ID 0x%04x.\n", device);
if (!(gpu_description = wined3d_get_gpu_description(vendor, device)))
{
ERR("Card %04x:%04x not found in driver DB.\n", vendor, device);
return FALSE;
}
}
fixup_extensions(gl_info, caps_gl_ctx, gl_renderer_str, gl_vendor,
gpu_description->vendor, gpu_description->device);
wined3d_driver_info_init(driver_info, gpu_description, vram_bytes, 0);
TRACE("Reporting (fake) driver version 0x%08x-0x%08x.\n",
driver_info->version_high, driver_info->version_low);
adapter->vram_bytes_used = 0;
TRACE("Emulating 0x%s bytes of video ram.\n", wine_dbgstr_longlong(driver_info->vram_bytes));
if (gl_info->supported[EXT_MEMORY_OBJECT])
{
GLint device_count = 0;
gl_info->gl_ops.gl.p_glGetIntegerv(GL_NUM_DEVICE_UUIDS_EXT, &device_count);
if (device_count > 0)
{
if (device_count > 1)
FIXME("A set of %d devices is not supported.\n", device_count);
GL_EXTCALL(glGetUnsignedBytevEXT(GL_DRIVER_UUID_EXT, (GLubyte *)&adapter->driver_uuid));
GL_EXTCALL(glGetUnsignedBytei_vEXT(GL_DEVICE_UUID_EXT, 0, (GLubyte *)&adapter->device_uuid));
TRACE("Driver UUID: %s, device UUID %s.\n",
debugstr_guid(&adapter->driver_uuid), debugstr_guid(&adapter->device_uuid));
}
else
{
WARN("Unexpected device count %d.\n", device_count);
}
}
gl_ext_emul_mask = adapter->vertex_pipe->vp_get_emul_mask(gl_info)
| adapter->fragment_pipe->get_emul_mask(gl_info);
if (gl_ext_emul_mask & GL_EXT_EMUL_ARB_MULTITEXTURE)
install_gl_compat_wrapper(gl_info, ARB_MULTITEXTURE);
if (gl_ext_emul_mask & GL_EXT_EMUL_EXT_FOG_COORD)
install_gl_compat_wrapper(gl_info, EXT_FOG_COORD);
return TRUE;
}
static void WINE_GLAPI invalid_func(const void *data)
{
ERR("Invalid vertex attribute function called.\n");
DebugBreak();
}
static void WINE_GLAPI invalid_texcoord_func(GLenum unit, const void *data)
{
ERR("Invalid texcoord function called.\n");
DebugBreak();
}
static void WINE_GLAPI invalid_generic_attrib_func(GLuint idx, const void *data)
{
ERR("Invalid attribute function called.\n");
DebugBreak();
}
/* Helper functions for providing vertex data to OpenGL. The arrays are
* initialised based on the extension detection and are used in
* draw_primitive_immediate_mode(). */
static void WINE_GLAPI position_d3dcolor(const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
DWORD pos = *((const DWORD *)data);
FIXME("Add a test for fixed function position from d3dcolor type.\n");
gl_info->gl_ops.gl.p_glVertex4s(D3DCOLOR_B_R(pos),
D3DCOLOR_B_G(pos),
D3DCOLOR_B_B(pos),
D3DCOLOR_B_A(pos));
}
static void WINE_GLAPI position_float4(const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
const GLfloat *pos = data;
if (pos[3] != 0.0f && pos[3] != 1.0f)
{
float w = 1.0f / pos[3];
gl_info->gl_ops.gl.p_glVertex4f(pos[0] * w, pos[1] * w, pos[2] * w, w);
}
else
{
gl_info->gl_ops.gl.p_glVertex3fv(pos);
}
}
static void WINE_GLAPI diffuse_d3dcolor(const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
DWORD diffuseColor = *((const DWORD *)data);
gl_info->gl_ops.gl.p_glColor4ub(D3DCOLOR_B_R(diffuseColor),
D3DCOLOR_B_G(diffuseColor),
D3DCOLOR_B_B(diffuseColor),
D3DCOLOR_B_A(diffuseColor));
}
static void WINE_GLAPI specular_d3dcolor(const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
DWORD specularColor = *((const DWORD *)data);
GLubyte d[] =
{
D3DCOLOR_B_R(specularColor),
D3DCOLOR_B_G(specularColor),
D3DCOLOR_B_B(specularColor)
};
gl_info->gl_ops.ext.p_glSecondaryColor3ubvEXT(d);
}
static void WINE_GLAPI warn_no_specular_func(const void *data)
{
WARN("GL_EXT_secondary_color not supported.\n");
}
static void WINE_GLAPI generic_d3dcolor(GLuint idx, const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
DWORD color = *((const DWORD *)data);
gl_info->gl_ops.ext.p_glVertexAttrib4Nub(idx,
D3DCOLOR_B_R(color), D3DCOLOR_B_G(color),
D3DCOLOR_B_B(color), D3DCOLOR_B_A(color));
}
static void WINE_GLAPI generic_short2n(GLuint idx, const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
const GLshort s[] = {((const GLshort *)data)[0], ((const GLshort *)data)[1], 0, 1};
gl_info->gl_ops.ext.p_glVertexAttrib4Nsv(idx, s);
}
static void WINE_GLAPI generic_ushort2n(GLuint idx, const void *data)
{
const GLushort s[] = {((const GLushort *)data)[0], ((const GLushort *)data)[1], 0, 1};
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
gl_info->gl_ops.ext.p_glVertexAttrib4Nusv(idx, s);
}
static void WINE_GLAPI generic_float16_2(GLuint idx, const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
float x = float_16_to_32(((const unsigned short *)data) + 0);
float y = float_16_to_32(((const unsigned short *)data) + 1);
gl_info->gl_ops.ext.p_glVertexAttrib2f(idx, x, y);
}
static void WINE_GLAPI generic_float16_4(GLuint idx, const void *data)
{
const struct wined3d_gl_info *gl_info = wined3d_context_gl_get_current()->gl_info;
float x = float_16_to_32(((const unsigned short *)data) + 0);
float y = float_16_to_32(((const unsigned short *)data) + 1);
float z = float_16_to_32(((const unsigned short *)data) + 2);
float w = float_16_to_32(((const unsigned short *)data) + 3);
gl_info->gl_ops.ext.p_glVertexAttrib4f(idx, x, y, z, w);
}
static void wined3d_adapter_init_ffp_attrib_ops(struct wined3d_adapter *adapter)
{
const struct wined3d_gl_info *gl_info = &adapter->gl_info;
struct wined3d_d3d_info *d3d_info = &adapter->d3d_info;
struct wined3d_ffp_attrib_ops *ops = &d3d_info->ffp_attrib_ops;
unsigned int i;
for (i = 0; i < WINED3D_FFP_EMIT_COUNT; ++i)
{
ops->position[i] = invalid_func;
ops->diffuse[i] = invalid_func;
ops->specular[i] = invalid_func;
ops->normal[i] = invalid_func;
ops->texcoord[i] = invalid_texcoord_func;
ops->generic[i] = invalid_generic_attrib_func;
}
ops->position[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glVertex3fv;
if (!d3d_info->xyzrhw)
ops->position[WINED3D_FFP_EMIT_FLOAT4] = position_float4;
else
ops->position[WINED3D_FFP_EMIT_FLOAT4] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glVertex4fv;
ops->position[WINED3D_FFP_EMIT_D3DCOLOR] = position_d3dcolor;
ops->position[WINED3D_FFP_EMIT_SHORT4] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glVertex2sv;
ops->diffuse[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glColor3fv;
ops->diffuse[WINED3D_FFP_EMIT_FLOAT4] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glColor4fv;
ops->diffuse[WINED3D_FFP_EMIT_D3DCOLOR] = diffuse_d3dcolor;
ops->diffuse[WINED3D_FFP_EMIT_UBYTE4N] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glColor4ubv;
ops->diffuse[WINED3D_FFP_EMIT_SHORT4N] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glColor4sv;
ops->diffuse[WINED3D_FFP_EMIT_USHORT4N] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glColor4usv;
/* No 4 component entry points here. */
if (gl_info->supported[EXT_SECONDARY_COLOR])
ops->specular[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_ffp_attrib_func)GL_EXTCALL(glSecondaryColor3fvEXT);
else
ops->specular[WINED3D_FFP_EMIT_FLOAT3] = warn_no_specular_func;
if (gl_info->supported[EXT_SECONDARY_COLOR])
ops->specular[WINED3D_FFP_EMIT_D3DCOLOR] = specular_d3dcolor;
else
ops->specular[WINED3D_FFP_EMIT_D3DCOLOR] = warn_no_specular_func;
/* Only 3 component entry points here. Test how others behave. Float4
* normals are used by one of our tests, trying to pass it to the pixel
* shader, which fails on Windows. */
ops->normal[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glNormal3fv;
/* Just ignore the 4th value. */
ops->normal[WINED3D_FFP_EMIT_FLOAT4] = (wined3d_ffp_attrib_func)gl_info->gl_ops.gl.p_glNormal3fv;
ops->texcoord[WINED3D_FFP_EMIT_FLOAT1] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord1fvARB;
ops->texcoord[WINED3D_FFP_EMIT_FLOAT2] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord2fvARB;
ops->texcoord[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord3fvARB;
ops->texcoord[WINED3D_FFP_EMIT_FLOAT4] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord4fvARB;
ops->texcoord[WINED3D_FFP_EMIT_SHORT2] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord2svARB;
ops->texcoord[WINED3D_FFP_EMIT_SHORT4] = (wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord4svARB;
if (gl_info->supported[NV_HALF_FLOAT])
{
/* Not supported by ARB_HALF_FLOAT_VERTEX, so check for NV_HALF_FLOAT. */
ops->texcoord[WINED3D_FFP_EMIT_FLOAT16_2] =
(wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord2hvNV;
ops->texcoord[WINED3D_FFP_EMIT_FLOAT16_4] =
(wined3d_ffp_texcoord_func)gl_info->gl_ops.ext.p_glMultiTexCoord4hvNV;
}
ops->generic[WINED3D_FFP_EMIT_FLOAT1] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib1fv;
ops->generic[WINED3D_FFP_EMIT_FLOAT2] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib2fv;
ops->generic[WINED3D_FFP_EMIT_FLOAT3] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib3fv;
ops->generic[WINED3D_FFP_EMIT_FLOAT4] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4fv;
if (gl_info->supported[ARB_VERTEX_ARRAY_BGRA])
ops->generic[WINED3D_FFP_EMIT_D3DCOLOR] = generic_d3dcolor;
else
ops->generic[WINED3D_FFP_EMIT_D3DCOLOR] =
(wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4Nubv;
ops->generic[WINED3D_FFP_EMIT_UBYTE4] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4ubv;
ops->generic[WINED3D_FFP_EMIT_SHORT2] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib2sv;
ops->generic[WINED3D_FFP_EMIT_SHORT4] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4sv;
ops->generic[WINED3D_FFP_EMIT_UBYTE4N] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4Nubv;
ops->generic[WINED3D_FFP_EMIT_SHORT2N] = generic_short2n;
ops->generic[WINED3D_FFP_EMIT_SHORT4N] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4Nsv;
ops->generic[WINED3D_FFP_EMIT_USHORT2N] = generic_ushort2n;
ops->generic[WINED3D_FFP_EMIT_USHORT4N] = (wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4Nusv;
if (gl_info->supported[NV_HALF_FLOAT] && gl_info->supported[NV_VERTEX_PROGRAM])
{
ops->generic[WINED3D_FFP_EMIT_FLOAT16_2] =
(wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib2hvNV;
ops->generic[WINED3D_FFP_EMIT_FLOAT16_4] =
(wined3d_generic_attrib_func)gl_info->gl_ops.ext.p_glVertexAttrib4hvNV;
}
else
{
ops->generic[WINED3D_FFP_EMIT_FLOAT16_2] = generic_float16_2;
ops->generic[WINED3D_FFP_EMIT_FLOAT16_4] = generic_float16_4;
}
}
static void wined3d_adapter_init_fb_cfgs(struct wined3d_adapter_gl *adapter_gl, HDC dc)
{
const struct wined3d_gl_info *gl_info = &adapter_gl->a.gl_info;
int i;
if (gl_info->supported[WGL_ARB_PIXEL_FORMAT])
{
UINT attrib_count = 0;
GLint cfg_count;
int attribs[11];
int values[11];
int attribute;
attribute = WGL_NUMBER_PIXEL_FORMATS_ARB;
GL_EXTCALL(wglGetPixelFormatAttribivARB(dc, 0, 0, 1, &attribute, &cfg_count));
adapter_gl->pixel_formats = heap_calloc(cfg_count, sizeof(*adapter_gl->pixel_formats));
attribs[attrib_count++] = WGL_RED_BITS_ARB;
attribs[attrib_count++] = WGL_GREEN_BITS_ARB;
attribs[attrib_count++] = WGL_BLUE_BITS_ARB;
attribs[attrib_count++] = WGL_ALPHA_BITS_ARB;
attribs[attrib_count++] = WGL_COLOR_BITS_ARB;
attribs[attrib_count++] = WGL_DEPTH_BITS_ARB;
attribs[attrib_count++] = WGL_STENCIL_BITS_ARB;
attribs[attrib_count++] = WGL_DRAW_TO_WINDOW_ARB;
attribs[attrib_count++] = WGL_PIXEL_TYPE_ARB;
attribs[attrib_count++] = WGL_DOUBLE_BUFFER_ARB;
attribs[attrib_count++] = WGL_AUX_BUFFERS_ARB;
for (i = 0, adapter_gl->pixel_format_count = 0; i < cfg_count; ++i)
{
struct wined3d_pixel_format *cfg = &adapter_gl->pixel_formats[adapter_gl->pixel_format_count];
int format_id = i + 1;
if (!GL_EXTCALL(wglGetPixelFormatAttribivARB(dc, format_id, 0, attrib_count, attribs, values)))
continue;
cfg->iPixelFormat = format_id;
cfg->redSize = values[0];
cfg->greenSize = values[1];
cfg->blueSize = values[2];
cfg->alphaSize = values[3];
cfg->colorSize = values[4];
cfg->depthSize = values[5];
cfg->stencilSize = values[6];
cfg->windowDrawable = values[7];
cfg->iPixelType = values[8];
cfg->doubleBuffer = values[9];
cfg->auxBuffers = values[10];
cfg->numSamples = 0;
/* Check multisample support. */
if (gl_info->supported[ARB_MULTISAMPLE])
{
int attribs[2] = {WGL_SAMPLE_BUFFERS_ARB, WGL_SAMPLES_ARB};
int values[2];
if (GL_EXTCALL(wglGetPixelFormatAttribivARB(dc, format_id, 0, 2, attribs, values)))
{
/* values[0] = WGL_SAMPLE_BUFFERS_ARB which tells whether
* multisampling is supported. values[1] = number of
* multisample buffers. */
if (values[0])
cfg->numSamples = values[1];
}
}
TRACE("iPixelFormat=%d, iPixelType=%#x, doubleBuffer=%d, RGBA=%d/%d/%d/%d, "
"depth=%d, stencil=%d, samples=%d, windowDrawable=%d\n",
cfg->iPixelFormat, cfg->iPixelType, cfg->doubleBuffer,
cfg->redSize, cfg->greenSize, cfg->blueSize, cfg->alphaSize,
cfg->depthSize, cfg->stencilSize, cfg->numSamples, cfg->windowDrawable);
++adapter_gl->pixel_format_count;
}
}
else
{
int cfg_count;
cfg_count = DescribePixelFormat(dc, 0, 0, 0);
adapter_gl->pixel_formats = heap_calloc(cfg_count, sizeof(*adapter_gl->pixel_formats));
for (i = 0, adapter_gl->pixel_format_count = 0; i < cfg_count; ++i)
{
struct wined3d_pixel_format *cfg = &adapter_gl->pixel_formats[adapter_gl->pixel_format_count];
PIXELFORMATDESCRIPTOR pfd;
int format_id = i + 1;
if (!DescribePixelFormat(dc, format_id, sizeof(pfd), &pfd))
continue;
/* We only want HW acceleration using an OpenGL ICD driver.
* PFD_GENERIC_FORMAT = slow OpenGL 1.1 GDI software rendering.
* PFD_GENERIC_ACCELERATED = partial hw acceleration using a MCD
* driver (e.g. 3dfx minigl). */
if (pfd.dwFlags & (PFD_GENERIC_FORMAT | PFD_GENERIC_ACCELERATED))
{
TRACE("Skipping format %d because it isn't ICD accelerated.\n", format_id);
continue;
}
cfg->iPixelFormat = format_id;
cfg->redSize = pfd.cRedBits;
cfg->greenSize = pfd.cGreenBits;
cfg->blueSize = pfd.cBlueBits;
cfg->alphaSize = pfd.cAlphaBits;
cfg->colorSize = pfd.cColorBits;
cfg->depthSize = pfd.cDepthBits;
cfg->stencilSize = pfd.cStencilBits;
cfg->windowDrawable = (pfd.dwFlags & PFD_DRAW_TO_WINDOW) ? 1 : 0;
cfg->iPixelType = (pfd.iPixelType == PFD_TYPE_RGBA) ? WGL_TYPE_RGBA_ARB : WGL_TYPE_COLORINDEX_ARB;
cfg->doubleBuffer = (pfd.dwFlags & PFD_DOUBLEBUFFER) ? 1 : 0;
cfg->auxBuffers = pfd.cAuxBuffers;
cfg->numSamples = 0;
TRACE("iPixelFormat=%d, iPixelType=%#x, doubleBuffer=%d, RGBA=%d/%d/%d/%d, "
"depth=%d, stencil=%d, windowDrawable=%d\n",
cfg->iPixelFormat, cfg->iPixelType, cfg->doubleBuffer,
cfg->redSize, cfg->greenSize, cfg->blueSize, cfg->alphaSize,
cfg->depthSize, cfg->stencilSize, cfg->windowDrawable);
++adapter_gl->pixel_format_count;
}
}
}
static void adapter_gl_destroy(struct wined3d_adapter *adapter)
{
struct wined3d_adapter_gl *adapter_gl = wined3d_adapter_gl(adapter);
heap_free(adapter_gl->pixel_formats);
wined3d_adapter_cleanup(adapter);
heap_free(adapter_gl);
}
static HRESULT adapter_gl_create_device(struct wined3d *wined3d, const struct wined3d_adapter *adapter,
enum wined3d_device_type device_type, HWND focus_window, unsigned int flags, BYTE surface_alignment,
const enum wined3d_feature_level *levels, unsigned int level_count,
struct wined3d_device_parent *device_parent, struct wined3d_device **device)
{
struct wined3d_device_gl *device_gl;
HRESULT hr;
if (!(device_gl = heap_alloc_zero(sizeof(*device_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_device_init(&device_gl->d, wined3d, adapter->ordinal, device_type,
focus_window, flags, surface_alignment, levels, level_count, device_parent)))
{
WARN("Failed to initialize device, hr %#x.\n", hr);
heap_free(device_gl);
return hr;
}
*device = &device_gl->d;
return WINED3D_OK;
}
static void adapter_gl_destroy_device(struct wined3d_device *device)
{
struct wined3d_device_gl *device_gl = wined3d_device_gl(device);
wined3d_device_cleanup(&device_gl->d);
heap_free(device_gl);
}
struct wined3d_context *adapter_gl_acquire_context(struct wined3d_device *device,
struct wined3d_texture *texture, unsigned int sub_resource_idx)
{
return wined3d_context_gl_acquire(device, texture, sub_resource_idx);
}
void adapter_gl_release_context(struct wined3d_context *context)
{
return wined3d_context_gl_release(wined3d_context_gl(context));
}
static void adapter_gl_get_wined3d_caps(const struct wined3d_adapter *adapter, struct wined3d_caps *caps)
{
const struct wined3d_d3d_info *d3d_info = &adapter->d3d_info;
const struct wined3d_gl_info *gl_info = &adapter->gl_info;
caps->ddraw_caps.dds_caps |= WINEDDSCAPS_BACKBUFFER
| WINEDDSCAPS_COMPLEX
| WINEDDSCAPS_FRONTBUFFER
| WINEDDSCAPS_3DDEVICE
| WINEDDSCAPS_VIDEOMEMORY
| WINEDDSCAPS_OWNDC
| WINEDDSCAPS_LOCALVIDMEM
| WINEDDSCAPS_NONLOCALVIDMEM;
caps->ddraw_caps.caps |= WINEDDCAPS_3D;
if (gl_info->supported[ARB_FRAMEBUFFER_OBJECT] || gl_info->supported[EXT_FRAMEBUFFER_OBJECT])
caps->Caps2 |= WINED3DCAPS2_CANGENMIPMAP;
if (gl_info->supported[WINED3D_GL_BLEND_EQUATION])
caps->PrimitiveMiscCaps |= WINED3DPMISCCAPS_BLENDOP;
if (gl_info->supported[EXT_BLEND_EQUATION_SEPARATE] && gl_info->supported[EXT_BLEND_FUNC_SEPARATE])
caps->PrimitiveMiscCaps |= WINED3DPMISCCAPS_SEPARATEALPHABLEND;
if (gl_info->supported[EXT_DRAW_BUFFERS2])
caps->PrimitiveMiscCaps |= WINED3DPMISCCAPS_INDEPENDENTWRITEMASKS;
if (gl_info->supported[ARB_FRAMEBUFFER_SRGB])
caps->PrimitiveMiscCaps |= WINED3DPMISCCAPS_POSTBLENDSRGBCONVERT;
if (~gl_info->quirks & WINED3D_QUIRK_NO_INDEPENDENT_BIT_DEPTHS)
caps->PrimitiveMiscCaps |= WINED3DPMISCCAPS_MRTINDEPENDENTBITDEPTHS;
if (gl_info->supported[ARB_SAMPLER_OBJECTS] || gl_info->supported[EXT_TEXTURE_LOD_BIAS])
caps->RasterCaps |= WINED3DPRASTERCAPS_MIPMAPLODBIAS;
if (gl_info->supported[ARB_TEXTURE_FILTER_ANISOTROPIC])
{
caps->RasterCaps |= WINED3DPRASTERCAPS_ANISOTROPY;
caps->TextureFilterCaps |= WINED3DPTFILTERCAPS_MAGFANISOTROPIC
| WINED3DPTFILTERCAPS_MINFANISOTROPIC;
}
if (gl_info->supported[ARB_BLEND_FUNC_EXTENDED])
caps->DestBlendCaps |= WINED3DPBLENDCAPS_SRCALPHASAT;
if (gl_info->supported[EXT_BLEND_COLOR])
{
caps->SrcBlendCaps |= WINED3DPBLENDCAPS_BLENDFACTOR;
caps->DestBlendCaps |= WINED3DPBLENDCAPS_BLENDFACTOR;
}
if (gl_info->supported[EXT_TEXTURE3D])
{
caps->TextureCaps |= WINED3DPTEXTURECAPS_VOLUMEMAP
| WINED3DPTEXTURECAPS_MIPVOLUMEMAP;
if (!d3d_info->texture_npot)
caps->TextureCaps |= WINED3DPTEXTURECAPS_VOLUMEMAP_POW2;
caps->VolumeTextureFilterCaps |= WINED3DPTFILTERCAPS_MAGFLINEAR
| WINED3DPTFILTERCAPS_MAGFPOINT
| WINED3DPTFILTERCAPS_MINFLINEAR
| WINED3DPTFILTERCAPS_MINFPOINT
| WINED3DPTFILTERCAPS_MIPFLINEAR
| WINED3DPTFILTERCAPS_MIPFPOINT
| WINED3DPTFILTERCAPS_LINEAR
| WINED3DPTFILTERCAPS_LINEARMIPLINEAR
| WINED3DPTFILTERCAPS_LINEARMIPNEAREST
| WINED3DPTFILTERCAPS_MIPLINEAR
| WINED3DPTFILTERCAPS_MIPNEAREST
| WINED3DPTFILTERCAPS_NEAREST;
caps->VolumeTextureAddressCaps |= WINED3DPTADDRESSCAPS_INDEPENDENTUV
| WINED3DPTADDRESSCAPS_CLAMP
| WINED3DPTADDRESSCAPS_WRAP;
if (gl_info->supported[ARB_TEXTURE_BORDER_CLAMP])
{
caps->VolumeTextureAddressCaps |= WINED3DPTADDRESSCAPS_BORDER;
}
if (gl_info->supported[ARB_TEXTURE_MIRRORED_REPEAT])
{
caps->VolumeTextureAddressCaps |= WINED3DPTADDRESSCAPS_MIRROR;
}
if (gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE])
{
caps->VolumeTextureAddressCaps |= WINED3DPTADDRESSCAPS_MIRRORONCE;
}
caps->MaxVolumeExtent = gl_info->limits.texture3d_size;
}
if (gl_info->supported[ARB_TEXTURE_CUBE_MAP])
{
caps->TextureCaps |= WINED3DPTEXTURECAPS_CUBEMAP
| WINED3DPTEXTURECAPS_MIPCUBEMAP;
if (!d3d_info->texture_npot)
caps->TextureCaps |= WINED3DPTEXTURECAPS_CUBEMAP_POW2;
caps->CubeTextureFilterCaps |= WINED3DPTFILTERCAPS_MAGFLINEAR
| WINED3DPTFILTERCAPS_MAGFPOINT
| WINED3DPTFILTERCAPS_MINFLINEAR
| WINED3DPTFILTERCAPS_MINFPOINT
| WINED3DPTFILTERCAPS_MIPFLINEAR
| WINED3DPTFILTERCAPS_MIPFPOINT
| WINED3DPTFILTERCAPS_LINEAR
| WINED3DPTFILTERCAPS_LINEARMIPLINEAR
| WINED3DPTFILTERCAPS_LINEARMIPNEAREST
| WINED3DPTFILTERCAPS_MIPLINEAR
| WINED3DPTFILTERCAPS_MIPNEAREST
| WINED3DPTFILTERCAPS_NEAREST;
if (gl_info->supported[ARB_TEXTURE_FILTER_ANISOTROPIC])
{
caps->CubeTextureFilterCaps |= WINED3DPTFILTERCAPS_MAGFANISOTROPIC
| WINED3DPTFILTERCAPS_MINFANISOTROPIC;
}
}
if (gl_info->supported[ARB_TEXTURE_BORDER_CLAMP])
{
caps->TextureAddressCaps |= WINED3DPTADDRESSCAPS_BORDER;
}
if (gl_info->supported[ARB_TEXTURE_MIRRORED_REPEAT])
{
caps->TextureAddressCaps |= WINED3DPTADDRESSCAPS_MIRROR;
}
if (gl_info->supported[ARB_TEXTURE_MIRROR_CLAMP_TO_EDGE])
{
caps->TextureAddressCaps |= WINED3DPTADDRESSCAPS_MIRRORONCE;
}
if (gl_info->supported[EXT_STENCIL_WRAP])
{
caps->StencilCaps |= WINED3DSTENCILCAPS_DECR
| WINED3DSTENCILCAPS_INCR;
}
if (gl_info->supported[WINED3D_GL_VERSION_2_0]
|| gl_info->supported[EXT_STENCIL_TWO_SIDE]
|| gl_info->supported[ATI_SEPARATE_STENCIL])
{
caps->StencilCaps |= WINED3DSTENCILCAPS_TWOSIDED;
}
caps->MaxAnisotropy = gl_info->limits.anisotropy;
if (caps->VertexShaderVersion >= 3)
{
caps->MaxVertexShader30InstructionSlots
= max(caps->MaxVertexShader30InstructionSlots, gl_info->limits.arb_vs_instructions);
}
if (caps->VertexShaderVersion >= 2)
{
caps->VS20Caps.temp_count = max(caps->VS20Caps.temp_count, gl_info->limits.arb_vs_temps);
if (gl_info->supported[ARB_HALF_FLOAT_VERTEX])
caps->DeclTypes |= WINED3DDTCAPS_FLOAT16_2 | WINED3DDTCAPS_FLOAT16_4;
}
if (caps->PixelShaderVersion >= 3)
{
caps->MaxPixelShader30InstructionSlots
= max(caps->MaxPixelShader30InstructionSlots, gl_info->limits.arb_ps_instructions);
}
if (caps->PixelShaderVersion >= 2)
{
caps->PS20Caps.temp_count = max(caps->PS20Caps.temp_count, gl_info->limits.arb_ps_temps);
}
}
static BOOL wined3d_check_pixel_format_color(const struct wined3d_pixel_format *cfg,
const struct wined3d_format *format)
{
/* Float formats need FBOs. If FBOs are used this function isn't called */
if (format->flags[WINED3D_GL_RES_TYPE_TEX_2D] & WINED3DFMT_FLAG_FLOAT)
return FALSE;
/* Probably a RGBA_float or color index mode. */
if (cfg->iPixelType != WGL_TYPE_RGBA_ARB)
return FALSE;
if (cfg->redSize < format->red_size
|| cfg->greenSize < format->green_size
|| cfg->blueSize < format->blue_size
|| cfg->alphaSize < format->alpha_size)
return FALSE;
return TRUE;
}
static BOOL wined3d_check_pixel_format_depth(const struct wined3d_pixel_format *cfg,
const struct wined3d_format *format)
{
BOOL lockable = FALSE;
/* Float formats need FBOs. If FBOs are used this function isn't called */
if (format->flags[WINED3D_GL_RES_TYPE_TEX_2D] & WINED3DFMT_FLAG_FLOAT)
return FALSE;
if ((format->id == WINED3DFMT_D16_LOCKABLE) || (format->id == WINED3DFMT_D32_FLOAT))
lockable = TRUE;
/* On some modern cards like the Geforce8/9, GLX doesn't offer some
* depth/stencil formats which D3D9 reports. We can safely report
* "compatible" formats (e.g. D24 can be used for D16) as long as we
* aren't dealing with a lockable format. This also helps D3D <= 7 as they
* expect D16 which isn't offered without this on Geforce8 cards. */
if (!(cfg->depthSize == format->depth_size || (!lockable && cfg->depthSize > format->depth_size)))
return FALSE;
/* Some cards like Intel i915 ones only offer D24S8 but lots of games also
* need a format without stencil. We can allow a mismatch if the format
* doesn't have any stencil bits. If it does have stencil bits the size
* must match, or stencil wrapping would break. */
if (format->stencil_size && cfg->stencilSize != format->stencil_size)
return FALSE;
return TRUE;
}
static BOOL adapter_gl_check_format(const struct wined3d_adapter *adapter,
const struct wined3d_format *adapter_format, const struct wined3d_format *rt_format,
const struct wined3d_format *ds_format)
{
const struct wined3d_adapter_gl *adapter_gl = wined3d_adapter_gl_const(adapter);
unsigned int i;
if (wined3d_settings.offscreen_rendering_mode != ORM_BACKBUFFER)
return TRUE;
if (adapter_format && rt_format)
{
/* In backbuffer mode the front and backbuffer share the same WGL
* pixelformat. The format must match in RGB, alpha is allowed to be
* different. (Only the backbuffer can have alpha.) */
if (adapter_format->red_size != rt_format->red_size
|| adapter_format->green_size != rt_format->green_size
|| adapter_format->blue_size != rt_format->blue_size)
{
TRACE("Render target format %s doesn't match with adapter format %s.\n",
debug_d3dformat(rt_format->id), debug_d3dformat(adapter_format->id));
return FALSE;
}
}
for (i = 0; i < adapter_gl->pixel_format_count; ++i)
{
const struct wined3d_pixel_format *cfg = &adapter_gl->pixel_formats[i];
/* Check if there is a WGL pixel format matching the requirements, the format should also be window
* drawable (not offscreen; e.g. Nvidia offers R5G6B5 for pbuffers even when X is running at 24bit) */
if (adapter_format && rt_format && !cfg->windowDrawable)
continue;
if ((!adapter_format || wined3d_check_pixel_format_color(cfg, adapter_format))
&& (!rt_format || wined3d_check_pixel_format_color(cfg, rt_format))
&& (!ds_format || wined3d_check_pixel_format_depth(cfg, ds_format)))
{
TRACE("Pixel format %d is compatible.\n", cfg->iPixelFormat);
return TRUE;
}
}
return FALSE;
}
static HRESULT adapter_gl_init_3d(struct wined3d_device *device)
{
TRACE("device %p.\n", device);
wined3d_cs_init_object(device->cs, wined3d_device_create_primary_opengl_context_cs, device);
wined3d_cs_finish(device->cs, WINED3D_CS_QUEUE_DEFAULT);
if (!wined3d_swapchain_gl(device->swapchains[0])->context_count)
return E_FAIL;
device->d3d_initialized = TRUE;
return WINED3D_OK;
}
static void adapter_gl_uninit_3d(struct wined3d_device *device)
{
TRACE("device %p.\n", device);
wined3d_cs_destroy_object(device->cs, wined3d_device_delete_opengl_contexts_cs, device);
wined3d_cs_finish(device->cs, WINED3D_CS_QUEUE_DEFAULT);
}
static void *adapter_gl_map_bo_address(struct wined3d_context *context,
const struct wined3d_bo_address *data, size_t size, uint32_t bind_flags, uint32_t map_flags)
{
struct wined3d_context_gl *context_gl;
GLenum binding;
context_gl = wined3d_context_gl(context);
binding = wined3d_buffer_gl_binding_from_bind_flags(context_gl->gl_info, bind_flags);
return wined3d_context_gl_map_bo_address(context_gl, data, size, binding, map_flags);
}
static void adapter_gl_unmap_bo_address(struct wined3d_context *context, const struct wined3d_bo_address *data,
uint32_t bind_flags, unsigned int range_count, const struct wined3d_map_range *ranges)
{
struct wined3d_context_gl *context_gl;
GLenum binding;
context_gl = wined3d_context_gl(context);
binding = wined3d_buffer_gl_binding_from_bind_flags(context_gl->gl_info, bind_flags);
wined3d_context_gl_unmap_bo_address(context_gl, data, binding, range_count, ranges);
}
static void adapter_gl_copy_bo_address(struct wined3d_context *context,
const struct wined3d_bo_address *dst, uint32_t dst_bind_flags,
const struct wined3d_bo_address *src, uint32_t src_bind_flags, size_t size)
{
struct wined3d_context_gl *context_gl;
GLenum dst_binding, src_binding;
context_gl = wined3d_context_gl(context);
dst_binding = wined3d_buffer_gl_binding_from_bind_flags(context_gl->gl_info, dst_bind_flags);
src_binding = wined3d_buffer_gl_binding_from_bind_flags(context_gl->gl_info, src_bind_flags);
wined3d_context_gl_copy_bo_address(context_gl, dst, dst_binding, src, src_binding, size);
}
static HRESULT adapter_gl_create_swapchain(struct wined3d_device *device, struct wined3d_swapchain_desc *desc,
void *parent, const struct wined3d_parent_ops *parent_ops, struct wined3d_swapchain **swapchain)
{
struct wined3d_swapchain_gl *swapchain_gl;
HRESULT hr;
TRACE("device %p, desc %p, parent %p, parent_ops %p, swapchain %p.\n",
device, desc, parent, parent_ops, swapchain);
if (!(swapchain_gl = heap_alloc_zero(sizeof(*swapchain_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_swapchain_gl_init(swapchain_gl, device, desc, parent, parent_ops)))
{
WARN("Failed to initialise swapchain, hr %#x.\n", hr);
heap_free(swapchain_gl);
return hr;
}
TRACE("Created swapchain %p.\n", swapchain_gl);
*swapchain = &swapchain_gl->s;
return hr;
}
static void adapter_gl_destroy_swapchain(struct wined3d_swapchain *swapchain)
{
struct wined3d_swapchain_gl *swapchain_gl = wined3d_swapchain_gl(swapchain);
wined3d_swapchain_gl_cleanup(swapchain_gl);
heap_free(swapchain_gl);
}
static HRESULT adapter_gl_create_buffer(struct wined3d_device *device,
const struct wined3d_buffer_desc *desc, const struct wined3d_sub_resource_data *data,
void *parent, const struct wined3d_parent_ops *parent_ops, struct wined3d_buffer **buffer)
{
struct wined3d_buffer_gl *buffer_gl;
HRESULT hr;
TRACE("device %p, desc %p, data %p, parent %p, parent_ops %p, buffer %p.\n",
device, desc, data, parent, parent_ops, buffer);
if (!(buffer_gl = heap_alloc_zero(sizeof(*buffer_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_buffer_gl_init(buffer_gl, device, desc, data, parent, parent_ops)))
{
WARN("Failed to initialise buffer, hr %#x.\n", hr);
heap_free(buffer_gl);
return hr;
}
TRACE("Created buffer %p.\n", buffer_gl);
*buffer = &buffer_gl->b;
return hr;
}
static void wined3d_buffer_gl_destroy_object(void *object)
{
struct wined3d_buffer_gl *buffer_gl = object;
struct wined3d_context *context;
if (buffer_gl->b.buffer_object)
{
context = context_acquire(buffer_gl->b.resource.device, NULL, 0);
wined3d_buffer_gl_destroy_buffer_object(buffer_gl, wined3d_context_gl(context));
context_release(context);
}
heap_free(buffer_gl);
}
static void adapter_gl_destroy_buffer(struct wined3d_buffer *buffer)
{
struct wined3d_buffer_gl *buffer_gl = wined3d_buffer_gl(buffer);
struct wined3d_device *device = buffer_gl->b.resource.device;
unsigned int swapchain_count = device->swapchain_count;
TRACE("buffer_gl %p.\n", buffer_gl);
/* Take a reference to the device, in case releasing the buffer would
* cause the device to be destroyed. However, swapchain resources don't
* take a reference to the device, and we wouldn't want to increment the
* refcount on a device that's in the process of being destroyed. */
if (swapchain_count)
wined3d_device_incref(device);
wined3d_buffer_cleanup(&buffer_gl->b);
wined3d_cs_destroy_object(device->cs, wined3d_buffer_gl_destroy_object, buffer_gl);
if (swapchain_count)
wined3d_device_decref(device);
}
static HRESULT adapter_gl_create_texture(struct wined3d_device *device,
const struct wined3d_resource_desc *desc, unsigned int layer_count, unsigned int level_count,
uint32_t flags, void *parent, const struct wined3d_parent_ops *parent_ops, struct wined3d_texture **texture)
{
struct wined3d_texture_gl *texture_gl;
HRESULT hr;
TRACE("device %p, desc %p, layer_count %u, level_count %u, flags %#x, parent %p, parent_ops %p, texture %p.\n",
device, desc, layer_count, level_count, flags, parent, parent_ops, texture);
if (!(texture_gl = wined3d_texture_allocate_object_memory(sizeof(*texture_gl), level_count, layer_count)))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_texture_gl_init(texture_gl, device, desc,
layer_count, level_count, flags, parent, parent_ops)))
{
WARN("Failed to initialise texture, hr %#x.\n", hr);
heap_free(texture_gl);
return hr;
}
TRACE("Created texture %p.\n", texture_gl);
*texture = &texture_gl->t;
return hr;
}
static void wined3d_texture_gl_destroy_object(void *object)
{
struct wined3d_renderbuffer_entry *entry, *entry2;
struct wined3d_texture_gl *texture_gl = object;
const struct wined3d_gl_info *gl_info;
struct wined3d_context *context;
struct wined3d_device *device;
TRACE("texture_gl %p.\n", texture_gl);
if (!list_empty(&texture_gl->renderbuffers))
{
device = texture_gl->t.resource.device;
context = context_acquire(device, NULL, 0);
gl_info = wined3d_context_gl(context)->gl_info;
LIST_FOR_EACH_ENTRY_SAFE(entry, entry2, &texture_gl->renderbuffers, struct wined3d_renderbuffer_entry, entry)
{
TRACE("Deleting renderbuffer %u.\n", entry->id);
context_gl_resource_released(device, entry->id, TRUE);
gl_info->fbo_ops.glDeleteRenderbuffers(1, &entry->id);
heap_free(entry);
}
context_release(context);
}
wined3d_texture_gl_unload_texture(texture_gl);
heap_free(texture_gl);
}
static void adapter_gl_destroy_texture(struct wined3d_texture *texture)
{
struct wined3d_texture_gl *texture_gl = wined3d_texture_gl(texture);
struct wined3d_device *device = texture_gl->t.resource.device;
unsigned int swapchain_count = device->swapchain_count;
TRACE("texture_gl %p.\n", texture_gl);
/* Take a reference to the device, in case releasing the texture would
* cause the device to be destroyed. However, swapchain resources don't
* take a reference to the device, and we wouldn't want to increment the
* refcount on a device that's in the process of being destroyed. */
if (swapchain_count)
wined3d_device_incref(device);
wined3d_texture_sub_resources_destroyed(texture);
texture->resource.parent_ops->wined3d_object_destroyed(texture->resource.parent);
wined3d_texture_cleanup(&texture_gl->t);
wined3d_cs_destroy_object(device->cs, wined3d_texture_gl_destroy_object, texture_gl);
if (swapchain_count)
wined3d_device_decref(device);
}
static HRESULT adapter_gl_create_rendertarget_view(const struct wined3d_view_desc *desc,
struct wined3d_resource *resource, void *parent, const struct wined3d_parent_ops *parent_ops,
struct wined3d_rendertarget_view **view)
{
struct wined3d_rendertarget_view_gl *view_gl;
HRESULT hr;
TRACE("desc %s, resource %p, parent %p, parent_ops %p, view %p.\n",
wined3d_debug_view_desc(desc, resource), resource, parent, parent_ops, view);
if (!(view_gl = heap_alloc_zero(sizeof(*view_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_rendertarget_view_gl_init(view_gl, desc, resource, parent, parent_ops)))
{
WARN("Failed to initialise view, hr %#x.\n", hr);
heap_free(view_gl);
return hr;
}
TRACE("Created render target view %p.\n", view_gl);
*view = &view_gl->v;
return hr;
}
struct wined3d_view_gl_destroy_ctx
{
struct wined3d_device *device;
const struct wined3d_gl_view *gl_view;
GLuint counter_bo;
void *object;
struct wined3d_view_gl_destroy_ctx *free;
};
static void wined3d_view_gl_destroy_object(void *object)
{
struct wined3d_view_gl_destroy_ctx *ctx = object;
const struct wined3d_gl_info *gl_info;
struct wined3d_context *context;
struct wined3d_device *device;
device = ctx->device;
if (ctx->gl_view->name || ctx->counter_bo)
{
context = context_acquire(device, NULL, 0);
gl_info = wined3d_context_gl(context)->gl_info;
if (ctx->gl_view->name)
{
context_gl_resource_released(device, ctx->gl_view->name, FALSE);
gl_info->gl_ops.gl.p_glDeleteTextures(1, &ctx->gl_view->name);
}
if (ctx->counter_bo)
GL_EXTCALL(glDeleteBuffers(1, &ctx->counter_bo));
checkGLcall("delete resources");
context_release(context);
}
heap_free(ctx->object);
heap_free(ctx->free);
}
static void wined3d_view_gl_destroy(struct wined3d_device *device,
const struct wined3d_gl_view *gl_view, GLuint counter_bo, void *object)
{
struct wined3d_view_gl_destroy_ctx *ctx, c;
if (!(ctx = heap_alloc(sizeof(*ctx))))
ctx = &c;
ctx->device = device;
ctx->gl_view = gl_view;
ctx->counter_bo = counter_bo;
ctx->object = object;
ctx->free = ctx != &c ? ctx : NULL;
wined3d_cs_destroy_object(device->cs, wined3d_view_gl_destroy_object, ctx);
if (!ctx->free)
device->cs->ops->finish(device->cs, WINED3D_CS_QUEUE_DEFAULT);
}
static void adapter_gl_destroy_rendertarget_view(struct wined3d_rendertarget_view *view)
{
struct wined3d_rendertarget_view_gl *view_gl = wined3d_rendertarget_view_gl(view);
struct wined3d_device *device = view_gl->v.resource->device;
unsigned int swapchain_count = device->swapchain_count;
TRACE("view_gl %p.\n", view_gl);
/* Take a reference to the device, in case releasing the view's resource
* would cause the device to be destroyed. However, swapchain resources
* don't take a reference to the device, and we wouldn't want to increment
* the refcount on a device that's in the process of being destroyed. */
if (swapchain_count)
wined3d_device_incref(device);
wined3d_rendertarget_view_cleanup(&view_gl->v);
wined3d_view_gl_destroy(device, &view_gl->gl_view, 0, view_gl);
if (swapchain_count)
wined3d_device_decref(device);
}
static HRESULT adapter_gl_create_shader_resource_view(const struct wined3d_view_desc *desc,
struct wined3d_resource *resource, void *parent, const struct wined3d_parent_ops *parent_ops,
struct wined3d_shader_resource_view **view)
{
struct wined3d_shader_resource_view_gl *view_gl;
HRESULT hr;
TRACE("desc %s, resource %p, parent %p, parent_ops %p, view %p.\n",
wined3d_debug_view_desc(desc, resource), resource, parent, parent_ops, view);
if (!(view_gl = heap_alloc_zero(sizeof(*view_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_shader_resource_view_gl_init(view_gl, desc, resource, parent, parent_ops)))
{
WARN("Failed to initialise view, hr %#x.\n", hr);
heap_free(view_gl);
return hr;
}
TRACE("Created shader resource view %p.\n", view_gl);
*view = &view_gl->v;
return hr;
}
static void adapter_gl_destroy_shader_resource_view(struct wined3d_shader_resource_view *view)
{
struct wined3d_shader_resource_view_gl *view_gl = wined3d_shader_resource_view_gl(view);
struct wined3d_device *device = view_gl->v.resource->device;
unsigned int swapchain_count = device->swapchain_count;
TRACE("view_gl %p.\n", view_gl);
/* Take a reference to the device, in case releasing the view's resource
* would cause the device to be destroyed. However, swapchain resources
* don't take a reference to the device, and we wouldn't want to increment
* the refcount on a device that's in the process of being destroyed. */
if (swapchain_count)
wined3d_device_incref(device);
wined3d_shader_resource_view_cleanup(&view_gl->v);
wined3d_view_gl_destroy(device, &view_gl->gl_view, 0, view_gl);
if (swapchain_count)
wined3d_device_decref(device);
}
static HRESULT adapter_gl_create_unordered_access_view(const struct wined3d_view_desc *desc,
struct wined3d_resource *resource, void *parent, const struct wined3d_parent_ops *parent_ops,
struct wined3d_unordered_access_view **view)
{
struct wined3d_unordered_access_view_gl *view_gl;
HRESULT hr;
TRACE("desc %s, resource %p, parent %p, parent_ops %p, view %p.\n",
wined3d_debug_view_desc(desc, resource), resource, parent, parent_ops, view);
if (!(view_gl = heap_alloc_zero(sizeof(*view_gl))))
return E_OUTOFMEMORY;
if (FAILED(hr = wined3d_unordered_access_view_gl_init(view_gl, desc, resource, parent, parent_ops)))
{
WARN("Failed to initialise view, hr %#x.\n", hr);
heap_free(view_gl);
return hr;
}
TRACE("Created unordered access view %p.\n", view_gl);
*view = &view_gl->v;
return hr;
}
static void adapter_gl_destroy_unordered_access_view(struct wined3d_unordered_access_view *view)
{
struct wined3d_unordered_access_view_gl *view_gl = wined3d_unordered_access_view_gl(view);
struct wined3d_device *device = view_gl->v.resource->device;
unsigned int swapchain_count = device->swapchain_count;
TRACE("view_gl %p.\n", view_gl);
/* Take a reference to the device, in case releasing the view's resource
* would cause the device to be destroyed. However, swapchain resources
* don't take a reference to the device, and we wouldn't want to increment
* the refcount on a device that's in the process of being destroyed. */
if (swapchain_count)
wined3d_device_incref(device);
wined3d_unordered_access_view_cleanup(&view_gl->v);
wined3d_view_gl_destroy(device, &view_gl->gl_view, view_gl->counter_bo, view_gl);
if (swapchain_count)
wined3d_device_decref(device);
}
static HRESULT adapter_gl_create_sampler(struct wined3d_device *device, const struct wined3d_sampler_desc *desc,
void *parent, const struct wined3d_parent_ops *parent_ops, struct wined3d_sampler **sampler)
{
struct wined3d_sampler_gl *sampler_gl;
TRACE("device %p, desc %p, parent %p, parent_ops %p, sampler %p.\n",
device, desc, parent, parent_ops, sampler);
if (!(sampler_gl = heap_alloc_zero(sizeof(*sampler_gl))))
return E_OUTOFMEMORY;
wined3d_sampler_gl_init(sampler_gl, device, desc, parent, parent_ops);
TRACE("Created sampler %p.\n", sampler_gl);
*sampler = &sampler_gl->s;
return WINED3D_OK;
}
static void wined3d_sampler_gl_destroy_object(void *object)
{
struct wined3d_sampler_gl *sampler_gl = object;
const struct wined3d_gl_info *gl_info;
struct wined3d_context *context;
if (sampler_gl->name)
{
context = context_acquire(sampler_gl->s.device, NULL, 0);
gl_info = wined3d_context_gl(context)->gl_info;
GL_EXTCALL(glDeleteSamplers(1, &sampler_gl->name));
context_release(context);
}
heap_free(sampler_gl);
}
static void adapter_gl_destroy_sampler(struct wined3d_sampler *sampler)
{
struct wined3d_sampler_gl *sampler_gl = wined3d_sampler_gl(sampler);
TRACE("sampler_gl %p.\n", sampler_gl);
wined3d_cs_destroy_object(sampler->device->cs, wined3d_sampler_gl_destroy_object, sampler_gl);
}
static HRESULT adapter_gl_create_query(struct wined3d_device *device, enum wined3d_query_type type,
void *parent, const struct wined3d_parent_ops *parent_ops, struct wined3d_query **query)
{
TRACE("device %p, type %#x, parent %p, parent_ops %p, query %p.\n",
device, type, parent, parent_ops, query);
return wined3d_query_gl_create(device, type, parent, parent_ops, query);
}
static void wined3d_query_gl_destroy_object(void *object)
{
struct wined3d_query *query = object;
if (query->buffer_object)
{
struct wined3d_context *context;
context = context_acquire(query->device, NULL, 0);
wined3d_query_gl_destroy_buffer_object(wined3d_context_gl(context), query);
context_release(context);
}
/* Queries are specific to the GL context that created them. Not
* deleting the query will obviously leak it, but that's still better
* than potentially deleting a different query with the same id in this
* context, and (still) leaking the actual query. */
query->query_ops->query_destroy(query);
}
static void adapter_gl_destroy_query(struct wined3d_query *query)
{
TRACE("query %p.\n", query);
wined3d_cs_destroy_object(query->device->cs, wined3d_query_gl_destroy_object, query);
}
static void adapter_gl_flush_context(struct wined3d_context *context)
{
struct wined3d_context_gl *context_gl = wined3d_context_gl(context);
TRACE("context_gl %p.\n", context_gl);
if (context_gl->valid)
context_gl->gl_info->gl_ops.gl.p_glFlush();
}
void adapter_gl_clear_uav(struct wined3d_context *context,
struct wined3d_unordered_access_view *view, const struct wined3d_uvec4 *clear_value)
{
TRACE("context %p, view %p, clear_value %s.\n", context, view, debug_uvec4(clear_value));
wined3d_unordered_access_view_gl_clear_uint(wined3d_unordered_access_view_gl(view),
clear_value, wined3d_context_gl(context));
}
static const struct wined3d_adapter_ops wined3d_adapter_gl_ops =
{
adapter_gl_destroy,
adapter_gl_create_device,
adapter_gl_destroy_device,
adapter_gl_acquire_context,
adapter_gl_release_context,
adapter_gl_get_wined3d_caps,
adapter_gl_check_format,
adapter_gl_init_3d,
adapter_gl_uninit_3d,
adapter_gl_map_bo_address,
adapter_gl_unmap_bo_address,
adapter_gl_copy_bo_address,
adapter_gl_create_swapchain,
adapter_gl_destroy_swapchain,
adapter_gl_create_buffer,
adapter_gl_destroy_buffer,
adapter_gl_create_texture,
adapter_gl_destroy_texture,
adapter_gl_create_rendertarget_view,
adapter_gl_destroy_rendertarget_view,
adapter_gl_create_shader_resource_view,
adapter_gl_destroy_shader_resource_view,
adapter_gl_create_unordered_access_view,
adapter_gl_destroy_unordered_access_view,
adapter_gl_create_sampler,
adapter_gl_destroy_sampler,
adapter_gl_create_query,
adapter_gl_destroy_query,
adapter_gl_flush_context,
adapter_gl_clear_uav,
};
static void wined3d_adapter_gl_init_d3d_info(struct wined3d_adapter_gl *adapter_gl, uint32_t wined3d_creation_flags)
{
const struct wined3d_gl_info *gl_info = &adapter_gl->a.gl_info;
struct wined3d_d3d_info *d3d_info = &adapter_gl->a.d3d_info;
struct wined3d_vertex_caps vertex_caps;
struct fragment_caps fragment_caps;
struct shader_caps shader_caps;
GLfloat f[2];
adapter_gl->a.shader_backend->shader_get_caps(&adapter_gl->a, &shader_caps);
adapter_gl->a.vertex_pipe->vp_get_caps(&adapter_gl->a, &vertex_caps);
adapter_gl->a.fragment_pipe->get_caps(&adapter_gl->a, &fragment_caps);
d3d_info->limits.vs_version = shader_caps.vs_version;
d3d_info->limits.hs_version = shader_caps.hs_version;
d3d_info->limits.ds_version = shader_caps.ds_version;
d3d_info->limits.gs_version = shader_caps.gs_version;
d3d_info->limits.ps_version = shader_caps.ps_version;
d3d_info->limits.cs_version = shader_caps.cs_version;
d3d_info->limits.vs_uniform_count = shader_caps.vs_uniform_count;
d3d_info->limits.ps_uniform_count = shader_caps.ps_uniform_count;
d3d_info->limits.varying_count = shader_caps.varying_count;
d3d_info->limits.ffp_textures = fragment_caps.MaxSimultaneousTextures;
d3d_info->limits.ffp_blend_stages = fragment_caps.MaxTextureBlendStages;
TRACE("Max texture stages: %u.\n", d3d_info->limits.ffp_blend_stages);
d3d_info->limits.ffp_vertex_blend_matrices = vertex_caps.max_vertex_blend_matrices;
d3d_info->limits.active_light_count = vertex_caps.max_active_lights;
d3d_info->limits.max_rt_count = gl_info->limits.buffers;
d3d_info->limits.max_clip_distances = gl_info->limits.user_clip_distances;
d3d_info->limits.texture_size = gl_info->limits.texture_size;
gl_info->gl_ops.gl.p_glGetFloatv(gl_info->supported[WINED3D_GL_LEGACY_CONTEXT]
? GL_ALIASED_POINT_SIZE_RANGE : GL_POINT_SIZE_RANGE, f);
d3d_info->limits.pointsize_max = f[1];
TRACE("Maximum point size support - max point size %.8e.\n", f[1]);
d3d_info->wined3d_creation_flags = wined3d_creation_flags;
d3d_info->xyzrhw = vertex_caps.xyzrhw;
d3d_info->emulated_flatshading = vertex_caps.emulated_flatshading;
d3d_info->ffp_generic_attributes = vertex_caps.ffp_generic_attributes;
d3d_info->ffp_alpha_test = !!gl_info->supported[WINED3D_GL_LEGACY_CONTEXT];
d3d_info->vs_clipping = shader_caps.wined3d_caps & WINED3D_SHADER_CAP_VS_CLIPPING;
d3d_info->shader_color_key = !!(fragment_caps.wined3d_caps & WINED3D_FRAGMENT_CAP_COLOR_KEY);
d3d_info->shader_double_precision = !!(shader_caps.wined3d_caps & WINED3D_SHADER_CAP_DOUBLE_PRECISION);
d3d_info->shader_output_interpolation = !!(shader_caps.wined3d_caps & WINED3D_SHADER_CAP_OUTPUT_INTERPOLATION);
d3d_info->viewport_array_index_any_shader = !!gl_info->supported[ARB_SHADER_VIEWPORT_LAYER_ARRAY];
d3d_info->texture_npot = !!gl_info->supported[ARB_TEXTURE_NON_POWER_OF_TWO];
d3d_info->texture_npot_conditional = gl_info->supported[WINED3D_GL_NORMALIZED_TEXRECT]
|| gl_info->supported[ARB_TEXTURE_RECTANGLE];
d3d_info->draw_base_vertex_offset = !!gl_info->supported[ARB_DRAW_ELEMENTS_BASE_VERTEX];
d3d_info->vertex_bgra = !!gl_info->supported[ARB_VERTEX_ARRAY_BGRA];
d3d_info->texture_swizzle = !!gl_info->supported[ARB_TEXTURE_SWIZZLE];
d3d_info->srgb_read_control = !!gl_info->supported[EXT_TEXTURE_SRGB_DECODE];
d3d_info->srgb_write_control = !!gl_info->supported[ARB_FRAMEBUFFER_SRGB];
d3d_info->clip_control = !!gl_info->supported[ARB_CLIP_CONTROL];
d3d_info->full_ffp_varyings = !!(shader_caps.wined3d_caps & WINED3D_SHADER_CAP_FULL_FFP_VARYINGS);
d3d_info->feature_level = feature_level_from_caps(gl_info, &shader_caps, &fragment_caps);
if (gl_info->supported[ARB_TEXTURE_MULTISAMPLE])
d3d_info->multisample_draw_location = WINED3D_LOCATION_TEXTURE_RGB;
else
d3d_info->multisample_draw_location = WINED3D_LOCATION_RB_MULTISAMPLE;
}
static BOOL wined3d_adapter_gl_init(struct wined3d_adapter_gl *adapter_gl,
unsigned int ordinal, unsigned int wined3d_creation_flags)
{
static const DWORD supported_gl_versions[] =
{
MAKEDWORD_VERSION(4, 4),
MAKEDWORD_VERSION(3, 2),
MAKEDWORD_VERSION(1, 0),
};
struct wined3d_gl_info *gl_info = &adapter_gl->a.gl_info;
struct wined3d_caps_gl_ctx caps_gl_ctx = {0};
unsigned int i;
TRACE("adapter_gl %p, ordinal %u, wined3d_creation_flags %#x.\n",
adapter_gl, ordinal, wined3d_creation_flags);
if (!wined3d_adapter_init(&adapter_gl->a, ordinal, &wined3d_adapter_gl_ops))
return FALSE;
/* Dynamically load all GL core functions */
#ifdef USE_WIN32_OPENGL
{
HMODULE mod_gl = GetModuleHandleA("opengl32.dll");
#define USE_GL_FUNC(f) gl_info->gl_ops.gl.p_##f = (void *)GetProcAddress(mod_gl, #f);
ALL_WGL_FUNCS
#undef USE_GL_FUNC
gl_info->gl_ops.wgl.p_wglSwapBuffers = (void *)GetProcAddress(mod_gl, "wglSwapBuffers");
gl_info->gl_ops.wgl.p_wglGetPixelFormat = (void *)GetProcAddress(mod_gl, "wglGetPixelFormat");
}
#else
/* To bypass the opengl32 thunks retrieve functions from the WGL driver instead of opengl32 */
{
HDC hdc = GetDC( 0 );
const struct opengl_funcs *wgl_driver = __wine_get_wgl_driver( hdc, WINE_WGL_DRIVER_VERSION );
ReleaseDC( 0, hdc );
if (!wgl_driver || wgl_driver == (void *)-1) return FALSE;
gl_info->gl_ops.wgl = wgl_driver->wgl;
gl_info->gl_ops.gl = wgl_driver->gl;
}
#endif
gl_info->p_glEnableWINE = gl_info->gl_ops.gl.p_glEnable;
gl_info->p_glDisableWINE = gl_info->gl_ops.gl.p_glDisable;
if (!wined3d_caps_gl_ctx_create(&adapter_gl->a, &caps_gl_ctx))
{
ERR("Failed to get a GL context for adapter %p.\n", adapter_gl);
return FALSE;
}
for (i = 0; i < ARRAY_SIZE(supported_gl_versions); ++i)
{
if (supported_gl_versions[i] <= wined3d_settings.max_gl_version)
break;
}
if (i == ARRAY_SIZE(supported_gl_versions))
{
ERR_(winediag)("Requested invalid GL version %u.%u.\n",
wined3d_settings.max_gl_version >> 16, wined3d_settings.max_gl_version & 0xffff);
i = ARRAY_SIZE(supported_gl_versions) - 1;
}
for (; i < ARRAY_SIZE(supported_gl_versions); ++i)
{
gl_info->selected_gl_version = supported_gl_versions[i];
if (wined3d_caps_gl_ctx_create_attribs(&caps_gl_ctx, gl_info))
break;
WARN("Couldn't create an OpenGL %u.%u context, trying fallback to a lower version.\n",
supported_gl_versions[i] >> 16, supported_gl_versions[i] & 0xffff);
}
if (!wined3d_adapter_init_gl_caps(&adapter_gl->a, &caps_gl_ctx, wined3d_creation_flags))
{
ERR("Failed to initialize GL caps for adapter %p.\n", adapter_gl);
wined3d_caps_gl_ctx_destroy(&caps_gl_ctx);
return FALSE;
}
wined3d_adapter_gl_init_d3d_info(adapter_gl, wined3d_creation_flags);
if (!adapter_gl->a.d3d_info.shader_color_key)
{
/* We do not want to deal with re-creating immutable texture storage
* for colour-keying emulation. */
WARN("Disabling ARB_texture_storage because fragment pipe doesn't support colour-keying.\n");
gl_info->supported[ARB_TEXTURE_STORAGE] = FALSE;
}
if (wined3d_settings.offscreen_rendering_mode == ORM_BACKBUFFER)
ERR_(winediag)("You are using the backbuffer for offscreen rendering. "
"This is unsupported, and will be removed in a future version.\n");
wined3d_adapter_init_fb_cfgs(adapter_gl, caps_gl_ctx.dc);
/* We haven't found any suitable formats. This should only happen in
* case of GDI software rendering, which is pretty useless anyway. */
if (!adapter_gl->pixel_format_count)
{
WARN("No suitable pixel formats found.\n");
wined3d_caps_gl_ctx_destroy(&caps_gl_ctx);
heap_free(adapter_gl->pixel_formats);
return FALSE;
}
if (!wined3d_adapter_gl_init_format_info(&adapter_gl->a, &caps_gl_ctx))
{
ERR("Failed to initialize GL format info.\n");
wined3d_caps_gl_ctx_destroy(&caps_gl_ctx);
heap_free(adapter_gl->pixel_formats);
return FALSE;
}
wined3d_caps_gl_ctx_destroy(&caps_gl_ctx);
wined3d_adapter_init_ffp_attrib_ops(&adapter_gl->a);
return TRUE;
}
struct wined3d_adapter *wined3d_adapter_gl_create(unsigned int ordinal, unsigned int wined3d_creation_flags)
{
struct wined3d_adapter_gl *adapter;
if (!(adapter = heap_alloc_zero(sizeof(*adapter))))
return NULL;
if (!wined3d_adapter_gl_init(adapter, ordinal, wined3d_creation_flags))
{
heap_free(adapter);
return NULL;
}
TRACE("Created adapter %p.\n", adapter);
return &adapter->a;
}