Get the program binary type correctly or else ATI makes massive (~400MB) shader caches. Also, don't need the line in the PixelShaderGen.

This commit is contained in:
Ryan Houdek
2011-12-29 01:35:50 -06:00
parent 8a48b42e4c
commit 6a5b56d25f
3 changed files with 19 additions and 19 deletions

View File

@ -568,8 +568,6 @@ const char *GeneratePixelShaderCode(DSTALPHA_MODE dstAlphaMode, API_TYPE ApiType
if (g_ActiveConfig.backend_info.bSupportsGLSLATTRBind)
WRITE(p, "#extension GL_ARB_explicit_attrib_location : enable\n");
if (g_ActiveConfig.backend_info.bSupportsGLSLBlend)
WRITE(p, "#extension GL_ARB_blend_func_extended : enable\n");
// Silly differences
WRITE(p, "#define float2 vec2\n");
WRITE(p, "#define float3 vec3\n");