mirror of
https://gitlab.winehq.org/wine/vkd3d.git
synced 2024-11-21 16:46:41 -08:00
16d9a10dd5
Analog to d3dbc_compiler.
6595 lines
264 KiB
C
6595 lines
264 KiB
C
/*
|
|
* TPF (Direct3D shader models 4 and 5 bytecode) support
|
|
*
|
|
* Copyright 2008-2009 Henri Verbeet for CodeWeavers
|
|
* Copyright 2010 Rico Schüller
|
|
* Copyright 2017 Józef Kucia for CodeWeavers
|
|
* Copyright 2019-2020 Zebediah Figura for CodeWeavers
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#include "hlsl.h"
|
|
#include "vkd3d_shader_private.h"
|
|
#include "vkd3d_d3dcommon.h"
|
|
|
|
#define SM4_MAX_SRC_COUNT 6
|
|
#define SM4_MAX_DST_COUNT 2
|
|
|
|
STATIC_ASSERT(SM4_MAX_SRC_COUNT <= SPIRV_MAX_SRC_COUNT);
|
|
|
|
#define VKD3D_SM4_PS 0x0000u
|
|
#define VKD3D_SM4_VS 0x0001u
|
|
#define VKD3D_SM4_GS 0x0002u
|
|
#define VKD3D_SM5_HS 0x0003u
|
|
#define VKD3D_SM5_DS 0x0004u
|
|
#define VKD3D_SM5_CS 0x0005u
|
|
#define VKD3D_SM4_LIB 0xfff0u
|
|
|
|
#define VKD3D_SM4_INSTRUCTION_MODIFIER (0x1u << 31)
|
|
|
|
#define VKD3D_SM4_MODIFIER_MASK 0x3fu
|
|
|
|
#define VKD3D_SM5_MODIFIER_DATA_TYPE_SHIFT 6
|
|
#define VKD3D_SM5_MODIFIER_DATA_TYPE_MASK (0xffffu << VKD3D_SM5_MODIFIER_DATA_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM5_MODIFIER_RESOURCE_TYPE_SHIFT 6
|
|
#define VKD3D_SM5_MODIFIER_RESOURCE_TYPE_MASK (0xfu << VKD3D_SM5_MODIFIER_RESOURCE_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM5_MODIFIER_RESOURCE_STRIDE_SHIFT 11
|
|
#define VKD3D_SM5_MODIFIER_RESOURCE_STRIDE_MASK (0xfffu << VKD3D_SM5_MODIFIER_RESOURCE_STRIDE_SHIFT)
|
|
|
|
#define VKD3D_SM4_AOFFIMMI_U_SHIFT 9
|
|
#define VKD3D_SM4_AOFFIMMI_U_MASK (0xfu << VKD3D_SM4_AOFFIMMI_U_SHIFT)
|
|
#define VKD3D_SM4_AOFFIMMI_V_SHIFT 13
|
|
#define VKD3D_SM4_AOFFIMMI_V_MASK (0xfu << VKD3D_SM4_AOFFIMMI_V_SHIFT)
|
|
#define VKD3D_SM4_AOFFIMMI_W_SHIFT 17
|
|
#define VKD3D_SM4_AOFFIMMI_W_MASK (0xfu << VKD3D_SM4_AOFFIMMI_W_SHIFT)
|
|
|
|
#define VKD3D_SM4_INSTRUCTION_LENGTH_SHIFT 24
|
|
#define VKD3D_SM4_INSTRUCTION_LENGTH_MASK (0x1fu << VKD3D_SM4_INSTRUCTION_LENGTH_SHIFT)
|
|
|
|
#define VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT 11
|
|
#define VKD3D_SM4_INSTRUCTION_FLAGS_MASK (0x7u << VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT)
|
|
|
|
#define VKD3D_SM4_RESOURCE_TYPE_SHIFT 11
|
|
#define VKD3D_SM4_RESOURCE_TYPE_MASK (0xfu << VKD3D_SM4_RESOURCE_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_RESOURCE_SAMPLE_COUNT_SHIFT 16
|
|
#define VKD3D_SM4_RESOURCE_SAMPLE_COUNT_MASK (0xfu << VKD3D_SM4_RESOURCE_SAMPLE_COUNT_SHIFT)
|
|
|
|
#define VKD3D_SM4_PRIMITIVE_TYPE_SHIFT 11
|
|
#define VKD3D_SM4_PRIMITIVE_TYPE_MASK (0x3fu << VKD3D_SM4_PRIMITIVE_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_INDEX_TYPE_SHIFT 11
|
|
#define VKD3D_SM4_INDEX_TYPE_MASK (0x1u << VKD3D_SM4_INDEX_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_SAMPLER_MODE_SHIFT 11
|
|
#define VKD3D_SM4_SAMPLER_MODE_MASK (0xfu << VKD3D_SM4_SAMPLER_MODE_SHIFT)
|
|
|
|
#define VKD3D_SM4_SHADER_DATA_TYPE_SHIFT 11
|
|
#define VKD3D_SM4_SHADER_DATA_TYPE_MASK (0xfu << VKD3D_SM4_SHADER_DATA_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_INTERPOLATION_MODE_SHIFT 11
|
|
#define VKD3D_SM4_INTERPOLATION_MODE_MASK (0xfu << VKD3D_SM4_INTERPOLATION_MODE_SHIFT)
|
|
|
|
#define VKD3D_SM4_GLOBAL_FLAGS_SHIFT 11
|
|
#define VKD3D_SM4_GLOBAL_FLAGS_MASK (0xffu << VKD3D_SM4_GLOBAL_FLAGS_SHIFT)
|
|
|
|
#define VKD3D_SM5_PRECISE_SHIFT 19
|
|
#define VKD3D_SM5_PRECISE_MASK (0xfu << VKD3D_SM5_PRECISE_SHIFT)
|
|
|
|
#define VKD3D_SM5_CONTROL_POINT_COUNT_SHIFT 11
|
|
#define VKD3D_SM5_CONTROL_POINT_COUNT_MASK (0xffu << VKD3D_SM5_CONTROL_POINT_COUNT_SHIFT)
|
|
|
|
#define VKD3D_SM5_FP_ARRAY_SIZE_SHIFT 16
|
|
#define VKD3D_SM5_FP_TABLE_COUNT_MASK 0xffffu
|
|
|
|
#define VKD3D_SM5_UAV_FLAGS_SHIFT 15
|
|
#define VKD3D_SM5_UAV_FLAGS_MASK (0x1ffu << VKD3D_SM5_UAV_FLAGS_SHIFT)
|
|
|
|
#define VKD3D_SM5_SYNC_FLAGS_SHIFT 11
|
|
#define VKD3D_SM5_SYNC_FLAGS_MASK (0xffu << VKD3D_SM5_SYNC_FLAGS_SHIFT)
|
|
|
|
#define VKD3D_SM5_TESSELLATOR_SHIFT 11
|
|
#define VKD3D_SM5_TESSELLATOR_MASK (0xfu << VKD3D_SM5_TESSELLATOR_SHIFT)
|
|
|
|
#define VKD3D_SM4_OPCODE_MASK 0xff
|
|
|
|
#define VKD3D_SM4_EXTENDED_OPERAND (0x1u << 31)
|
|
|
|
#define VKD3D_SM4_EXTENDED_OPERAND_TYPE_MASK 0x3fu
|
|
|
|
#define VKD3D_SM4_REGISTER_MODIFIER_SHIFT 6
|
|
#define VKD3D_SM4_REGISTER_MODIFIER_MASK (0xffu << VKD3D_SM4_REGISTER_MODIFIER_SHIFT)
|
|
|
|
#define VKD3D_SM4_REGISTER_PRECISION_SHIFT 14
|
|
#define VKD3D_SM4_REGISTER_PRECISION_MASK (0x7u << VKD3D_SM4_REGISTER_PRECISION_SHIFT)
|
|
|
|
#define VKD3D_SM4_REGISTER_NON_UNIFORM_SHIFT 17
|
|
#define VKD3D_SM4_REGISTER_NON_UNIFORM_MASK (0x1u << VKD3D_SM4_REGISTER_NON_UNIFORM_SHIFT)
|
|
|
|
#define VKD3D_SM4_ADDRESSING_SHIFT2 28
|
|
#define VKD3D_SM4_ADDRESSING_MASK2 (0x3u << VKD3D_SM4_ADDRESSING_SHIFT2)
|
|
|
|
#define VKD3D_SM4_ADDRESSING_SHIFT1 25
|
|
#define VKD3D_SM4_ADDRESSING_MASK1 (0x3u << VKD3D_SM4_ADDRESSING_SHIFT1)
|
|
|
|
#define VKD3D_SM4_ADDRESSING_SHIFT0 22
|
|
#define VKD3D_SM4_ADDRESSING_MASK0 (0x3u << VKD3D_SM4_ADDRESSING_SHIFT0)
|
|
|
|
#define VKD3D_SM4_REGISTER_ORDER_SHIFT 20
|
|
#define VKD3D_SM4_REGISTER_ORDER_MASK (0x3u << VKD3D_SM4_REGISTER_ORDER_SHIFT)
|
|
|
|
#define VKD3D_SM4_REGISTER_TYPE_SHIFT 12
|
|
#define VKD3D_SM4_REGISTER_TYPE_MASK (0xffu << VKD3D_SM4_REGISTER_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_SWIZZLE_TYPE_SHIFT 2
|
|
#define VKD3D_SM4_SWIZZLE_TYPE_MASK (0x3u << VKD3D_SM4_SWIZZLE_TYPE_SHIFT)
|
|
|
|
#define VKD3D_SM4_DIMENSION_SHIFT 0
|
|
#define VKD3D_SM4_DIMENSION_MASK (0x3u << VKD3D_SM4_DIMENSION_SHIFT)
|
|
|
|
#define VKD3D_SM4_WRITEMASK_SHIFT 4
|
|
#define VKD3D_SM4_WRITEMASK_MASK (0xfu << VKD3D_SM4_WRITEMASK_SHIFT)
|
|
|
|
#define VKD3D_SM4_SWIZZLE_SHIFT 4
|
|
#define VKD3D_SM4_SWIZZLE_MASK (0xffu << VKD3D_SM4_SWIZZLE_SHIFT)
|
|
|
|
#define VKD3D_SM4_SCALAR_DIM_SHIFT 4
|
|
#define VKD3D_SM4_SCALAR_DIM_MASK (0x3u << VKD3D_SM4_SCALAR_DIM_SHIFT)
|
|
|
|
#define VKD3D_SM4_VERSION_MAJOR(version) (((version) >> 4) & 0xf)
|
|
#define VKD3D_SM4_VERSION_MINOR(version) (((version) >> 0) & 0xf)
|
|
|
|
#define VKD3D_SM4_ADDRESSING_RELATIVE 0x2
|
|
#define VKD3D_SM4_ADDRESSING_OFFSET 0x1
|
|
|
|
#define VKD3D_SM4_INSTRUCTION_FLAG_SATURATE 0x4
|
|
|
|
#define VKD3D_SM4_CONDITIONAL_NZ (0x1u << 18)
|
|
|
|
#define VKD3D_SM4_TYPE_COMPONENT(com, i) (((com) >> (4 * (i))) & 0xfu)
|
|
|
|
/* The shift that corresponds to the D3D_SIF_TEXTURE_COMPONENTS mask. */
|
|
#define VKD3D_SM4_SIF_TEXTURE_COMPONENTS_SHIFT 2
|
|
|
|
#define VKD3D_SM4_REQUIRES_DOUBLES 0x00000001
|
|
#define VKD3D_SM4_REQUIRES_EARLY_DEPTH_STENCIL 0x00000002
|
|
#define VKD3D_SM4_REQUIRES_UAVS_AT_EVERY_STAGE 0x00000004
|
|
#define VKD3D_SM4_REQUIRES_64_UAVS 0x00000008
|
|
#define VKD3D_SM4_REQUIRES_MINIMUM_PRECISION 0x00000010
|
|
#define VKD3D_SM4_REQUIRES_11_1_DOUBLE_EXTENSIONS 0x00000020
|
|
#define VKD3D_SM4_REQUIRES_11_1_SHADER_EXTENSIONS 0x00000040
|
|
#define VKD3D_SM4_REQUIRES_LEVEL_9_COMPARISON_FILTERING 0x00000080
|
|
#define VKD3D_SM4_REQUIRES_TILED_RESOURCES 0x00000100
|
|
#define VKD3D_SM4_REQUIRES_STENCIL_REF 0x00000200
|
|
#define VKD3D_SM4_REQUIRES_INNER_COVERAGE 0x00000400
|
|
#define VKD3D_SM4_REQUIRES_TYPED_UAV_LOAD_ADDITIONAL_FORMATS 0x00000800
|
|
#define VKD3D_SM4_REQUIRES_ROVS 0x00001000
|
|
#define VKD3D_SM4_REQUIRES_VIEWPORT_AND_RT_ARRAY_INDEX_FROM_ANY_SHADER_FEEDING_RASTERIZER 0x00002000
|
|
|
|
enum vkd3d_sm4_opcode
|
|
{
|
|
VKD3D_SM4_OP_ADD = 0x00,
|
|
VKD3D_SM4_OP_AND = 0x01,
|
|
VKD3D_SM4_OP_BREAK = 0x02,
|
|
VKD3D_SM4_OP_BREAKC = 0x03,
|
|
VKD3D_SM4_OP_CASE = 0x06,
|
|
VKD3D_SM4_OP_CONTINUE = 0x07,
|
|
VKD3D_SM4_OP_CONTINUEC = 0x08,
|
|
VKD3D_SM4_OP_CUT = 0x09,
|
|
VKD3D_SM4_OP_DEFAULT = 0x0a,
|
|
VKD3D_SM4_OP_DERIV_RTX = 0x0b,
|
|
VKD3D_SM4_OP_DERIV_RTY = 0x0c,
|
|
VKD3D_SM4_OP_DISCARD = 0x0d,
|
|
VKD3D_SM4_OP_DIV = 0x0e,
|
|
VKD3D_SM4_OP_DP2 = 0x0f,
|
|
VKD3D_SM4_OP_DP3 = 0x10,
|
|
VKD3D_SM4_OP_DP4 = 0x11,
|
|
VKD3D_SM4_OP_ELSE = 0x12,
|
|
VKD3D_SM4_OP_EMIT = 0x13,
|
|
VKD3D_SM4_OP_ENDIF = 0x15,
|
|
VKD3D_SM4_OP_ENDLOOP = 0x16,
|
|
VKD3D_SM4_OP_ENDSWITCH = 0x17,
|
|
VKD3D_SM4_OP_EQ = 0x18,
|
|
VKD3D_SM4_OP_EXP = 0x19,
|
|
VKD3D_SM4_OP_FRC = 0x1a,
|
|
VKD3D_SM4_OP_FTOI = 0x1b,
|
|
VKD3D_SM4_OP_FTOU = 0x1c,
|
|
VKD3D_SM4_OP_GE = 0x1d,
|
|
VKD3D_SM4_OP_IADD = 0x1e,
|
|
VKD3D_SM4_OP_IF = 0x1f,
|
|
VKD3D_SM4_OP_IEQ = 0x20,
|
|
VKD3D_SM4_OP_IGE = 0x21,
|
|
VKD3D_SM4_OP_ILT = 0x22,
|
|
VKD3D_SM4_OP_IMAD = 0x23,
|
|
VKD3D_SM4_OP_IMAX = 0x24,
|
|
VKD3D_SM4_OP_IMIN = 0x25,
|
|
VKD3D_SM4_OP_IMUL = 0x26,
|
|
VKD3D_SM4_OP_INE = 0x27,
|
|
VKD3D_SM4_OP_INEG = 0x28,
|
|
VKD3D_SM4_OP_ISHL = 0x29,
|
|
VKD3D_SM4_OP_ISHR = 0x2a,
|
|
VKD3D_SM4_OP_ITOF = 0x2b,
|
|
VKD3D_SM4_OP_LABEL = 0x2c,
|
|
VKD3D_SM4_OP_LD = 0x2d,
|
|
VKD3D_SM4_OP_LD2DMS = 0x2e,
|
|
VKD3D_SM4_OP_LOG = 0x2f,
|
|
VKD3D_SM4_OP_LOOP = 0x30,
|
|
VKD3D_SM4_OP_LT = 0x31,
|
|
VKD3D_SM4_OP_MAD = 0x32,
|
|
VKD3D_SM4_OP_MIN = 0x33,
|
|
VKD3D_SM4_OP_MAX = 0x34,
|
|
VKD3D_SM4_OP_SHADER_DATA = 0x35,
|
|
VKD3D_SM4_OP_MOV = 0x36,
|
|
VKD3D_SM4_OP_MOVC = 0x37,
|
|
VKD3D_SM4_OP_MUL = 0x38,
|
|
VKD3D_SM4_OP_NE = 0x39,
|
|
VKD3D_SM4_OP_NOP = 0x3a,
|
|
VKD3D_SM4_OP_NOT = 0x3b,
|
|
VKD3D_SM4_OP_OR = 0x3c,
|
|
VKD3D_SM4_OP_RESINFO = 0x3d,
|
|
VKD3D_SM4_OP_RET = 0x3e,
|
|
VKD3D_SM4_OP_RETC = 0x3f,
|
|
VKD3D_SM4_OP_ROUND_NE = 0x40,
|
|
VKD3D_SM4_OP_ROUND_NI = 0x41,
|
|
VKD3D_SM4_OP_ROUND_PI = 0x42,
|
|
VKD3D_SM4_OP_ROUND_Z = 0x43,
|
|
VKD3D_SM4_OP_RSQ = 0x44,
|
|
VKD3D_SM4_OP_SAMPLE = 0x45,
|
|
VKD3D_SM4_OP_SAMPLE_C = 0x46,
|
|
VKD3D_SM4_OP_SAMPLE_C_LZ = 0x47,
|
|
VKD3D_SM4_OP_SAMPLE_LOD = 0x48,
|
|
VKD3D_SM4_OP_SAMPLE_GRAD = 0x49,
|
|
VKD3D_SM4_OP_SAMPLE_B = 0x4a,
|
|
VKD3D_SM4_OP_SQRT = 0x4b,
|
|
VKD3D_SM4_OP_SWITCH = 0x4c,
|
|
VKD3D_SM4_OP_SINCOS = 0x4d,
|
|
VKD3D_SM4_OP_UDIV = 0x4e,
|
|
VKD3D_SM4_OP_ULT = 0x4f,
|
|
VKD3D_SM4_OP_UGE = 0x50,
|
|
VKD3D_SM4_OP_UMUL = 0x51,
|
|
VKD3D_SM4_OP_UMAX = 0x53,
|
|
VKD3D_SM4_OP_UMIN = 0x54,
|
|
VKD3D_SM4_OP_USHR = 0x55,
|
|
VKD3D_SM4_OP_UTOF = 0x56,
|
|
VKD3D_SM4_OP_XOR = 0x57,
|
|
VKD3D_SM4_OP_DCL_RESOURCE = 0x58,
|
|
VKD3D_SM4_OP_DCL_CONSTANT_BUFFER = 0x59,
|
|
VKD3D_SM4_OP_DCL_SAMPLER = 0x5a,
|
|
VKD3D_SM4_OP_DCL_INDEX_RANGE = 0x5b,
|
|
VKD3D_SM4_OP_DCL_OUTPUT_TOPOLOGY = 0x5c,
|
|
VKD3D_SM4_OP_DCL_INPUT_PRIMITIVE = 0x5d,
|
|
VKD3D_SM4_OP_DCL_VERTICES_OUT = 0x5e,
|
|
VKD3D_SM4_OP_DCL_INPUT = 0x5f,
|
|
VKD3D_SM4_OP_DCL_INPUT_SGV = 0x60,
|
|
VKD3D_SM4_OP_DCL_INPUT_SIV = 0x61,
|
|
VKD3D_SM4_OP_DCL_INPUT_PS = 0x62,
|
|
VKD3D_SM4_OP_DCL_INPUT_PS_SGV = 0x63,
|
|
VKD3D_SM4_OP_DCL_INPUT_PS_SIV = 0x64,
|
|
VKD3D_SM4_OP_DCL_OUTPUT = 0x65,
|
|
VKD3D_SM4_OP_DCL_OUTPUT_SIV = 0x67,
|
|
VKD3D_SM4_OP_DCL_TEMPS = 0x68,
|
|
VKD3D_SM4_OP_DCL_INDEXABLE_TEMP = 0x69,
|
|
VKD3D_SM4_OP_DCL_GLOBAL_FLAGS = 0x6a,
|
|
VKD3D_SM4_OP_LOD = 0x6c,
|
|
VKD3D_SM4_OP_GATHER4 = 0x6d,
|
|
VKD3D_SM4_OP_SAMPLE_POS = 0x6e,
|
|
VKD3D_SM4_OP_SAMPLE_INFO = 0x6f,
|
|
VKD3D_SM5_OP_HS_DECLS = 0x71,
|
|
VKD3D_SM5_OP_HS_CONTROL_POINT_PHASE = 0x72,
|
|
VKD3D_SM5_OP_HS_FORK_PHASE = 0x73,
|
|
VKD3D_SM5_OP_HS_JOIN_PHASE = 0x74,
|
|
VKD3D_SM5_OP_EMIT_STREAM = 0x75,
|
|
VKD3D_SM5_OP_CUT_STREAM = 0x76,
|
|
VKD3D_SM5_OP_FCALL = 0x78,
|
|
VKD3D_SM5_OP_BUFINFO = 0x79,
|
|
VKD3D_SM5_OP_DERIV_RTX_COARSE = 0x7a,
|
|
VKD3D_SM5_OP_DERIV_RTX_FINE = 0x7b,
|
|
VKD3D_SM5_OP_DERIV_RTY_COARSE = 0x7c,
|
|
VKD3D_SM5_OP_DERIV_RTY_FINE = 0x7d,
|
|
VKD3D_SM5_OP_GATHER4_C = 0x7e,
|
|
VKD3D_SM5_OP_GATHER4_PO = 0x7f,
|
|
VKD3D_SM5_OP_GATHER4_PO_C = 0x80,
|
|
VKD3D_SM5_OP_RCP = 0x81,
|
|
VKD3D_SM5_OP_F32TOF16 = 0x82,
|
|
VKD3D_SM5_OP_F16TOF32 = 0x83,
|
|
VKD3D_SM5_OP_COUNTBITS = 0x86,
|
|
VKD3D_SM5_OP_FIRSTBIT_HI = 0x87,
|
|
VKD3D_SM5_OP_FIRSTBIT_LO = 0x88,
|
|
VKD3D_SM5_OP_FIRSTBIT_SHI = 0x89,
|
|
VKD3D_SM5_OP_UBFE = 0x8a,
|
|
VKD3D_SM5_OP_IBFE = 0x8b,
|
|
VKD3D_SM5_OP_BFI = 0x8c,
|
|
VKD3D_SM5_OP_BFREV = 0x8d,
|
|
VKD3D_SM5_OP_SWAPC = 0x8e,
|
|
VKD3D_SM5_OP_DCL_STREAM = 0x8f,
|
|
VKD3D_SM5_OP_DCL_FUNCTION_BODY = 0x90,
|
|
VKD3D_SM5_OP_DCL_FUNCTION_TABLE = 0x91,
|
|
VKD3D_SM5_OP_DCL_INTERFACE = 0x92,
|
|
VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT = 0x93,
|
|
VKD3D_SM5_OP_DCL_OUTPUT_CONTROL_POINT_COUNT = 0x94,
|
|
VKD3D_SM5_OP_DCL_TESSELLATOR_DOMAIN = 0x95,
|
|
VKD3D_SM5_OP_DCL_TESSELLATOR_PARTITIONING = 0x96,
|
|
VKD3D_SM5_OP_DCL_TESSELLATOR_OUTPUT_PRIMITIVE = 0x97,
|
|
VKD3D_SM5_OP_DCL_HS_MAX_TESSFACTOR = 0x98,
|
|
VKD3D_SM5_OP_DCL_HS_FORK_PHASE_INSTANCE_COUNT = 0x99,
|
|
VKD3D_SM5_OP_DCL_HS_JOIN_PHASE_INSTANCE_COUNT = 0x9a,
|
|
VKD3D_SM5_OP_DCL_THREAD_GROUP = 0x9b,
|
|
VKD3D_SM5_OP_DCL_UAV_TYPED = 0x9c,
|
|
VKD3D_SM5_OP_DCL_UAV_RAW = 0x9d,
|
|
VKD3D_SM5_OP_DCL_UAV_STRUCTURED = 0x9e,
|
|
VKD3D_SM5_OP_DCL_TGSM_RAW = 0x9f,
|
|
VKD3D_SM5_OP_DCL_TGSM_STRUCTURED = 0xa0,
|
|
VKD3D_SM5_OP_DCL_RESOURCE_RAW = 0xa1,
|
|
VKD3D_SM5_OP_DCL_RESOURCE_STRUCTURED = 0xa2,
|
|
VKD3D_SM5_OP_LD_UAV_TYPED = 0xa3,
|
|
VKD3D_SM5_OP_STORE_UAV_TYPED = 0xa4,
|
|
VKD3D_SM5_OP_LD_RAW = 0xa5,
|
|
VKD3D_SM5_OP_STORE_RAW = 0xa6,
|
|
VKD3D_SM5_OP_LD_STRUCTURED = 0xa7,
|
|
VKD3D_SM5_OP_STORE_STRUCTURED = 0xa8,
|
|
VKD3D_SM5_OP_ATOMIC_AND = 0xa9,
|
|
VKD3D_SM5_OP_ATOMIC_OR = 0xaa,
|
|
VKD3D_SM5_OP_ATOMIC_XOR = 0xab,
|
|
VKD3D_SM5_OP_ATOMIC_CMP_STORE = 0xac,
|
|
VKD3D_SM5_OP_ATOMIC_IADD = 0xad,
|
|
VKD3D_SM5_OP_ATOMIC_IMAX = 0xae,
|
|
VKD3D_SM5_OP_ATOMIC_IMIN = 0xaf,
|
|
VKD3D_SM5_OP_ATOMIC_UMAX = 0xb0,
|
|
VKD3D_SM5_OP_ATOMIC_UMIN = 0xb1,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_ALLOC = 0xb2,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_CONSUME = 0xb3,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_IADD = 0xb4,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_AND = 0xb5,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_OR = 0xb6,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_XOR = 0xb7,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_EXCH = 0xb8,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_CMP_EXCH = 0xb9,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_IMAX = 0xba,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_IMIN = 0xbb,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_UMAX = 0xbc,
|
|
VKD3D_SM5_OP_IMM_ATOMIC_UMIN = 0xbd,
|
|
VKD3D_SM5_OP_SYNC = 0xbe,
|
|
VKD3D_SM5_OP_DADD = 0xbf,
|
|
VKD3D_SM5_OP_DMAX = 0xc0,
|
|
VKD3D_SM5_OP_DMIN = 0xc1,
|
|
VKD3D_SM5_OP_DMUL = 0xc2,
|
|
VKD3D_SM5_OP_DEQ = 0xc3,
|
|
VKD3D_SM5_OP_DGE = 0xc4,
|
|
VKD3D_SM5_OP_DLT = 0xc5,
|
|
VKD3D_SM5_OP_DNE = 0xc6,
|
|
VKD3D_SM5_OP_DMOV = 0xc7,
|
|
VKD3D_SM5_OP_DMOVC = 0xc8,
|
|
VKD3D_SM5_OP_DTOF = 0xc9,
|
|
VKD3D_SM5_OP_FTOD = 0xca,
|
|
VKD3D_SM5_OP_EVAL_SAMPLE_INDEX = 0xcc,
|
|
VKD3D_SM5_OP_EVAL_CENTROID = 0xcd,
|
|
VKD3D_SM5_OP_DCL_GS_INSTANCES = 0xce,
|
|
VKD3D_SM5_OP_DDIV = 0xd2,
|
|
VKD3D_SM5_OP_DFMA = 0xd3,
|
|
VKD3D_SM5_OP_DRCP = 0xd4,
|
|
VKD3D_SM5_OP_MSAD = 0xd5,
|
|
VKD3D_SM5_OP_DTOI = 0xd6,
|
|
VKD3D_SM5_OP_DTOU = 0xd7,
|
|
VKD3D_SM5_OP_ITOD = 0xd8,
|
|
VKD3D_SM5_OP_UTOD = 0xd9,
|
|
VKD3D_SM5_OP_GATHER4_S = 0xdb,
|
|
VKD3D_SM5_OP_GATHER4_C_S = 0xdc,
|
|
VKD3D_SM5_OP_GATHER4_PO_S = 0xdd,
|
|
VKD3D_SM5_OP_GATHER4_PO_C_S = 0xde,
|
|
VKD3D_SM5_OP_LD_S = 0xdf,
|
|
VKD3D_SM5_OP_LD2DMS_S = 0xe0,
|
|
VKD3D_SM5_OP_LD_UAV_TYPED_S = 0xe1,
|
|
VKD3D_SM5_OP_LD_RAW_S = 0xe2,
|
|
VKD3D_SM5_OP_LD_STRUCTURED_S = 0xe3,
|
|
VKD3D_SM5_OP_SAMPLE_LOD_S = 0xe4,
|
|
VKD3D_SM5_OP_SAMPLE_C_LZ_S = 0xe5,
|
|
VKD3D_SM5_OP_SAMPLE_CL_S = 0xe6,
|
|
VKD3D_SM5_OP_SAMPLE_B_CL_S = 0xe7,
|
|
VKD3D_SM5_OP_SAMPLE_GRAD_CL_S = 0xe8,
|
|
VKD3D_SM5_OP_SAMPLE_C_CL_S = 0xe9,
|
|
VKD3D_SM5_OP_CHECK_ACCESS_FULLY_MAPPED = 0xea,
|
|
|
|
VKD3D_SM4_OP_COUNT,
|
|
};
|
|
|
|
enum vkd3d_sm4_instruction_modifier
|
|
{
|
|
VKD3D_SM4_MODIFIER_AOFFIMMI = 0x1,
|
|
VKD3D_SM5_MODIFIER_RESOURCE_TYPE = 0x2,
|
|
VKD3D_SM5_MODIFIER_DATA_TYPE = 0x3,
|
|
};
|
|
|
|
enum vkd3d_sm4_register_type
|
|
{
|
|
VKD3D_SM4_RT_TEMP = 0x00,
|
|
VKD3D_SM4_RT_INPUT = 0x01,
|
|
VKD3D_SM4_RT_OUTPUT = 0x02,
|
|
VKD3D_SM4_RT_INDEXABLE_TEMP = 0x03,
|
|
VKD3D_SM4_RT_IMMCONST = 0x04,
|
|
VKD3D_SM4_RT_IMMCONST64 = 0x05,
|
|
VKD3D_SM4_RT_SAMPLER = 0x06,
|
|
VKD3D_SM4_RT_RESOURCE = 0x07,
|
|
VKD3D_SM4_RT_CONSTBUFFER = 0x08,
|
|
VKD3D_SM4_RT_IMMCONSTBUFFER = 0x09,
|
|
VKD3D_SM4_RT_PRIMID = 0x0b,
|
|
VKD3D_SM4_RT_DEPTHOUT = 0x0c,
|
|
VKD3D_SM4_RT_NULL = 0x0d,
|
|
VKD3D_SM4_RT_RASTERIZER = 0x0e,
|
|
VKD3D_SM4_RT_OMASK = 0x0f,
|
|
VKD3D_SM5_RT_STREAM = 0x10,
|
|
VKD3D_SM5_RT_FUNCTION_BODY = 0x11,
|
|
VKD3D_SM5_RT_FUNCTION_POINTER = 0x13,
|
|
VKD3D_SM5_RT_OUTPUT_CONTROL_POINT_ID = 0x16,
|
|
VKD3D_SM5_RT_FORK_INSTANCE_ID = 0x17,
|
|
VKD3D_SM5_RT_JOIN_INSTANCE_ID = 0x18,
|
|
VKD3D_SM5_RT_INPUT_CONTROL_POINT = 0x19,
|
|
VKD3D_SM5_RT_OUTPUT_CONTROL_POINT = 0x1a,
|
|
VKD3D_SM5_RT_PATCH_CONSTANT_DATA = 0x1b,
|
|
VKD3D_SM5_RT_DOMAIN_LOCATION = 0x1c,
|
|
VKD3D_SM5_RT_UAV = 0x1e,
|
|
VKD3D_SM5_RT_SHARED_MEMORY = 0x1f,
|
|
VKD3D_SM5_RT_THREAD_ID = 0x20,
|
|
VKD3D_SM5_RT_THREAD_GROUP_ID = 0x21,
|
|
VKD3D_SM5_RT_LOCAL_THREAD_ID = 0x22,
|
|
VKD3D_SM5_RT_COVERAGE = 0x23,
|
|
VKD3D_SM5_RT_LOCAL_THREAD_INDEX = 0x24,
|
|
VKD3D_SM5_RT_GS_INSTANCE_ID = 0x25,
|
|
VKD3D_SM5_RT_DEPTHOUT_GREATER_EQUAL = 0x26,
|
|
VKD3D_SM5_RT_DEPTHOUT_LESS_EQUAL = 0x27,
|
|
VKD3D_SM5_RT_OUTPUT_STENCIL_REF = 0x29,
|
|
|
|
VKD3D_SM4_REGISTER_TYPE_COUNT,
|
|
};
|
|
|
|
enum vkd3d_sm4_extended_operand_type
|
|
{
|
|
VKD3D_SM4_EXTENDED_OPERAND_NONE = 0x0,
|
|
VKD3D_SM4_EXTENDED_OPERAND_MODIFIER = 0x1,
|
|
};
|
|
|
|
enum vkd3d_sm4_register_modifier
|
|
{
|
|
VKD3D_SM4_REGISTER_MODIFIER_NONE = 0x00,
|
|
VKD3D_SM4_REGISTER_MODIFIER_NEGATE = 0x01,
|
|
VKD3D_SM4_REGISTER_MODIFIER_ABS = 0x02,
|
|
VKD3D_SM4_REGISTER_MODIFIER_ABS_NEGATE = 0x03,
|
|
};
|
|
|
|
enum vkd3d_sm4_register_precision
|
|
{
|
|
VKD3D_SM4_REGISTER_PRECISION_DEFAULT = 0x0,
|
|
VKD3D_SM4_REGISTER_PRECISION_MIN_FLOAT_16 = 0x1,
|
|
VKD3D_SM4_REGISTER_PRECISION_MIN_FLOAT_10 = 0x2,
|
|
VKD3D_SM4_REGISTER_PRECISION_MIN_INT_16 = 0x4,
|
|
VKD3D_SM4_REGISTER_PRECISION_MIN_UINT_16 = 0x5,
|
|
};
|
|
|
|
enum vkd3d_sm4_output_primitive_type
|
|
{
|
|
VKD3D_SM4_OUTPUT_PT_POINTLIST = 0x1,
|
|
VKD3D_SM4_OUTPUT_PT_LINESTRIP = 0x3,
|
|
VKD3D_SM4_OUTPUT_PT_TRIANGLESTRIP = 0x5,
|
|
};
|
|
|
|
enum vkd3d_sm4_input_primitive_type
|
|
{
|
|
VKD3D_SM4_INPUT_PT_POINT = 0x01,
|
|
VKD3D_SM4_INPUT_PT_LINE = 0x02,
|
|
VKD3D_SM4_INPUT_PT_TRIANGLE = 0x03,
|
|
VKD3D_SM4_INPUT_PT_LINEADJ = 0x06,
|
|
VKD3D_SM4_INPUT_PT_TRIANGLEADJ = 0x07,
|
|
VKD3D_SM5_INPUT_PT_PATCH1 = 0x08,
|
|
VKD3D_SM5_INPUT_PT_PATCH2 = 0x09,
|
|
VKD3D_SM5_INPUT_PT_PATCH3 = 0x0a,
|
|
VKD3D_SM5_INPUT_PT_PATCH4 = 0x0b,
|
|
VKD3D_SM5_INPUT_PT_PATCH5 = 0x0c,
|
|
VKD3D_SM5_INPUT_PT_PATCH6 = 0x0d,
|
|
VKD3D_SM5_INPUT_PT_PATCH7 = 0x0e,
|
|
VKD3D_SM5_INPUT_PT_PATCH8 = 0x0f,
|
|
VKD3D_SM5_INPUT_PT_PATCH9 = 0x10,
|
|
VKD3D_SM5_INPUT_PT_PATCH10 = 0x11,
|
|
VKD3D_SM5_INPUT_PT_PATCH11 = 0x12,
|
|
VKD3D_SM5_INPUT_PT_PATCH12 = 0x13,
|
|
VKD3D_SM5_INPUT_PT_PATCH13 = 0x14,
|
|
VKD3D_SM5_INPUT_PT_PATCH14 = 0x15,
|
|
VKD3D_SM5_INPUT_PT_PATCH15 = 0x16,
|
|
VKD3D_SM5_INPUT_PT_PATCH16 = 0x17,
|
|
VKD3D_SM5_INPUT_PT_PATCH17 = 0x18,
|
|
VKD3D_SM5_INPUT_PT_PATCH18 = 0x19,
|
|
VKD3D_SM5_INPUT_PT_PATCH19 = 0x1a,
|
|
VKD3D_SM5_INPUT_PT_PATCH20 = 0x1b,
|
|
VKD3D_SM5_INPUT_PT_PATCH21 = 0x1c,
|
|
VKD3D_SM5_INPUT_PT_PATCH22 = 0x1d,
|
|
VKD3D_SM5_INPUT_PT_PATCH23 = 0x1e,
|
|
VKD3D_SM5_INPUT_PT_PATCH24 = 0x1f,
|
|
VKD3D_SM5_INPUT_PT_PATCH25 = 0x20,
|
|
VKD3D_SM5_INPUT_PT_PATCH26 = 0x21,
|
|
VKD3D_SM5_INPUT_PT_PATCH27 = 0x22,
|
|
VKD3D_SM5_INPUT_PT_PATCH28 = 0x23,
|
|
VKD3D_SM5_INPUT_PT_PATCH29 = 0x24,
|
|
VKD3D_SM5_INPUT_PT_PATCH30 = 0x25,
|
|
VKD3D_SM5_INPUT_PT_PATCH31 = 0x26,
|
|
VKD3D_SM5_INPUT_PT_PATCH32 = 0x27,
|
|
};
|
|
|
|
enum vkd3d_sm4_swizzle_type
|
|
{
|
|
VKD3D_SM4_SWIZZLE_NONE = 0x0, /* swizzle bitfield contains a mask */
|
|
VKD3D_SM4_SWIZZLE_VEC4 = 0x1,
|
|
VKD3D_SM4_SWIZZLE_SCALAR = 0x2,
|
|
|
|
VKD3D_SM4_SWIZZLE_DEFAULT = ~0u - 1,
|
|
VKD3D_SM4_SWIZZLE_INVALID = ~0u,
|
|
};
|
|
|
|
enum vkd3d_sm4_dimension
|
|
{
|
|
VKD3D_SM4_DIMENSION_NONE = 0x0,
|
|
VKD3D_SM4_DIMENSION_SCALAR = 0x1,
|
|
VKD3D_SM4_DIMENSION_VEC4 = 0x2,
|
|
};
|
|
|
|
static enum vsir_dimension vsir_dimension_from_sm4_dimension(enum vkd3d_sm4_dimension dim)
|
|
{
|
|
switch (dim)
|
|
{
|
|
case VKD3D_SM4_DIMENSION_NONE:
|
|
return VSIR_DIMENSION_NONE;
|
|
case VKD3D_SM4_DIMENSION_SCALAR:
|
|
return VSIR_DIMENSION_SCALAR;
|
|
case VKD3D_SM4_DIMENSION_VEC4:
|
|
return VSIR_DIMENSION_VEC4;
|
|
default:
|
|
FIXME("Unknown SM4 dimension %#x.\n", dim);
|
|
return VSIR_DIMENSION_NONE;
|
|
}
|
|
}
|
|
|
|
static enum vkd3d_sm4_dimension sm4_dimension_from_vsir_dimension(enum vsir_dimension dim)
|
|
{
|
|
switch (dim)
|
|
{
|
|
case VSIR_DIMENSION_NONE:
|
|
return VKD3D_SM4_DIMENSION_NONE;
|
|
case VSIR_DIMENSION_SCALAR:
|
|
return VKD3D_SM4_DIMENSION_SCALAR;
|
|
case VSIR_DIMENSION_VEC4:
|
|
return VKD3D_SM4_DIMENSION_VEC4;
|
|
case VSIR_DIMENSION_COUNT:
|
|
vkd3d_unreachable();
|
|
}
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
enum vkd3d_sm4_resource_type
|
|
{
|
|
VKD3D_SM4_RESOURCE_BUFFER = 0x1,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_1D = 0x2,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_2D = 0x3,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_2DMS = 0x4,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_3D = 0x5,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_CUBE = 0x6,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_1DARRAY = 0x7,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_2DARRAY = 0x8,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_2DMSARRAY = 0x9,
|
|
VKD3D_SM4_RESOURCE_TEXTURE_CUBEARRAY = 0xa,
|
|
VKD3D_SM4_RESOURCE_RAW_BUFFER = 0xb,
|
|
VKD3D_SM4_RESOURCE_STRUCTURED_BUFFER = 0xc,
|
|
};
|
|
|
|
enum vkd3d_sm4_data_type
|
|
{
|
|
VKD3D_SM4_DATA_UNORM = 0x1,
|
|
VKD3D_SM4_DATA_SNORM = 0x2,
|
|
VKD3D_SM4_DATA_INT = 0x3,
|
|
VKD3D_SM4_DATA_UINT = 0x4,
|
|
VKD3D_SM4_DATA_FLOAT = 0x5,
|
|
VKD3D_SM4_DATA_MIXED = 0x6,
|
|
VKD3D_SM4_DATA_DOUBLE = 0x7,
|
|
VKD3D_SM4_DATA_CONTINUED = 0x8,
|
|
VKD3D_SM4_DATA_UNUSED = 0x9,
|
|
};
|
|
|
|
enum vkd3d_sm4_sampler_mode
|
|
{
|
|
VKD3D_SM4_SAMPLER_DEFAULT = 0x0,
|
|
VKD3D_SM4_SAMPLER_COMPARISON = 0x1,
|
|
};
|
|
|
|
enum vkd3d_sm4_shader_data_type
|
|
{
|
|
VKD3D_SM4_SHADER_DATA_IMMEDIATE_CONSTANT_BUFFER = 0x3,
|
|
VKD3D_SM4_SHADER_DATA_MESSAGE = 0x4,
|
|
};
|
|
|
|
enum vkd3d_sm4_stat_field
|
|
{
|
|
VKD3D_STAT_UNUSED = 0,
|
|
VKD3D_STAT_INSTR_COUNT,
|
|
VKD3D_STAT_MOV,
|
|
VKD3D_STAT_MOVC,
|
|
VKD3D_STAT_CONV,
|
|
VKD3D_STAT_FLOAT,
|
|
VKD3D_STAT_INT,
|
|
VKD3D_STAT_UINT,
|
|
VKD3D_STAT_EMIT,
|
|
VKD3D_STAT_CUT,
|
|
VKD3D_STAT_SAMPLE,
|
|
VKD3D_STAT_SAMPLE_C,
|
|
VKD3D_STAT_SAMPLE_GRAD,
|
|
VKD3D_STAT_SAMPLE_BIAS,
|
|
VKD3D_STAT_LOAD,
|
|
VKD3D_STAT_STORE,
|
|
VKD3D_STAT_DCL_VERTICES_OUT,
|
|
VKD3D_STAT_DCL_INPUT_PRIMITIVE,
|
|
VKD3D_STAT_DCL_OUTPUT_TOPOLOGY,
|
|
VKD3D_STAT_DCL_GS_INSTANCES,
|
|
VKD3D_STAT_BITWISE,
|
|
VKD3D_STAT_ATOMIC,
|
|
VKD3D_STAT_TESS_DOMAIN,
|
|
VKD3D_STAT_TESS_PARTITIONING,
|
|
VKD3D_STAT_TESS_OUTPUT_PRIMITIVE,
|
|
VKD3D_STAT_TESS_CONTROL_POINT_COUNT,
|
|
VKD3D_STAT_BARRIER,
|
|
VKD3D_STAT_LOD,
|
|
VKD3D_STAT_GATHER,
|
|
VKD3D_STAT_TEMPS,
|
|
VKD3D_STAT_COUNT,
|
|
};
|
|
|
|
struct vkd3d_sm4_stat_field_info
|
|
{
|
|
enum vkd3d_sm4_opcode opcode;
|
|
enum vkd3d_sm4_stat_field field;
|
|
};
|
|
|
|
struct sm4_index_range
|
|
{
|
|
unsigned int index;
|
|
unsigned int count;
|
|
unsigned int mask;
|
|
};
|
|
|
|
struct sm4_index_range_array
|
|
{
|
|
unsigned int count;
|
|
struct sm4_index_range ranges[MAX_REG_OUTPUT * 2];
|
|
};
|
|
|
|
struct vkd3d_sm4_lookup_tables
|
|
{
|
|
const struct vkd3d_sm4_opcode_info *opcode_info_from_sm4[VKD3D_SM4_OP_COUNT];
|
|
const struct vkd3d_sm4_register_type_info *register_type_info_from_sm4[VKD3D_SM4_REGISTER_TYPE_COUNT];
|
|
const struct vkd3d_sm4_register_type_info *register_type_info_from_vkd3d[VKD3DSPR_COUNT];
|
|
const struct vkd3d_sm4_stat_field_info *stat_field_from_sm4[VKD3D_SM4_OP_COUNT];
|
|
};
|
|
|
|
struct vkd3d_shader_sm4_parser
|
|
{
|
|
const uint32_t *start, *end, *ptr;
|
|
|
|
enum vkd3d_shader_opcode phase;
|
|
bool has_control_point_phase;
|
|
unsigned int input_register_masks[MAX_REG_OUTPUT];
|
|
unsigned int output_register_masks[MAX_REG_OUTPUT];
|
|
unsigned int patch_constant_register_masks[MAX_REG_OUTPUT];
|
|
|
|
struct sm4_index_range_array input_index_ranges;
|
|
struct sm4_index_range_array output_index_ranges;
|
|
struct sm4_index_range_array patch_constant_index_ranges;
|
|
|
|
struct vkd3d_sm4_lookup_tables lookup;
|
|
|
|
struct vkd3d_shader_parser p;
|
|
};
|
|
|
|
struct vkd3d_sm4_opcode_info
|
|
{
|
|
enum vkd3d_sm4_opcode opcode;
|
|
enum vkd3d_shader_opcode handler_idx;
|
|
char dst_info[SM4_MAX_DST_COUNT];
|
|
char src_info[SM4_MAX_SRC_COUNT];
|
|
void (*read_opcode_func)(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv);
|
|
};
|
|
|
|
static const enum vkd3d_primitive_type output_primitive_type_table[] =
|
|
{
|
|
/* UNKNOWN */ VKD3D_PT_UNDEFINED,
|
|
/* VKD3D_SM4_OUTPUT_PT_POINTLIST */ VKD3D_PT_POINTLIST,
|
|
/* UNKNOWN */ VKD3D_PT_UNDEFINED,
|
|
/* VKD3D_SM4_OUTPUT_PT_LINESTRIP */ VKD3D_PT_LINESTRIP,
|
|
/* UNKNOWN */ VKD3D_PT_UNDEFINED,
|
|
/* VKD3D_SM4_OUTPUT_PT_TRIANGLESTRIP */ VKD3D_PT_TRIANGLESTRIP,
|
|
};
|
|
|
|
static const struct
|
|
{
|
|
unsigned int control_point_count;
|
|
enum vkd3d_primitive_type vkd3d_type;
|
|
}
|
|
input_primitive_type_table[] =
|
|
{
|
|
[VKD3D_SM4_INPUT_PT_POINT] = {1, VKD3D_PT_POINTLIST},
|
|
[VKD3D_SM4_INPUT_PT_LINE] = {2, VKD3D_PT_LINELIST},
|
|
[VKD3D_SM4_INPUT_PT_TRIANGLE] = {3, VKD3D_PT_TRIANGLELIST},
|
|
[VKD3D_SM4_INPUT_PT_LINEADJ] = {4, VKD3D_PT_LINELIST_ADJ},
|
|
[VKD3D_SM4_INPUT_PT_TRIANGLEADJ] = {6, VKD3D_PT_TRIANGLELIST_ADJ},
|
|
};
|
|
|
|
static const enum vkd3d_shader_resource_type resource_type_table[] =
|
|
{
|
|
/* 0 */ VKD3D_SHADER_RESOURCE_NONE,
|
|
/* VKD3D_SM4_RESOURCE_BUFFER */ VKD3D_SHADER_RESOURCE_BUFFER,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_1D */ VKD3D_SHADER_RESOURCE_TEXTURE_1D,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_2D */ VKD3D_SHADER_RESOURCE_TEXTURE_2D,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_2DMS */ VKD3D_SHADER_RESOURCE_TEXTURE_2DMS,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_3D */ VKD3D_SHADER_RESOURCE_TEXTURE_3D,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_CUBE */ VKD3D_SHADER_RESOURCE_TEXTURE_CUBE,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_1DARRAY */ VKD3D_SHADER_RESOURCE_TEXTURE_1DARRAY,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_2DARRAY */ VKD3D_SHADER_RESOURCE_TEXTURE_2DARRAY,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_2DMSARRAY */ VKD3D_SHADER_RESOURCE_TEXTURE_2DMSARRAY,
|
|
/* VKD3D_SM4_RESOURCE_TEXTURE_CUBEARRAY */ VKD3D_SHADER_RESOURCE_TEXTURE_CUBEARRAY,
|
|
/* VKD3D_SM4_RESOURCE_RAW_BUFFER */ VKD3D_SHADER_RESOURCE_BUFFER,
|
|
/* VKD3D_SM4_RESOURCE_STRUCTURED_BUFFER */ VKD3D_SHADER_RESOURCE_BUFFER,
|
|
};
|
|
|
|
static const enum vkd3d_data_type data_type_table[] =
|
|
{
|
|
/* 0 */ VKD3D_DATA_FLOAT,
|
|
/* VKD3D_SM4_DATA_UNORM */ VKD3D_DATA_UNORM,
|
|
/* VKD3D_SM4_DATA_SNORM */ VKD3D_DATA_SNORM,
|
|
/* VKD3D_SM4_DATA_INT */ VKD3D_DATA_INT,
|
|
/* VKD3D_SM4_DATA_UINT */ VKD3D_DATA_UINT,
|
|
/* VKD3D_SM4_DATA_FLOAT */ VKD3D_DATA_FLOAT,
|
|
/* VKD3D_SM4_DATA_MIXED */ VKD3D_DATA_MIXED,
|
|
/* VKD3D_SM4_DATA_DOUBLE */ VKD3D_DATA_DOUBLE,
|
|
/* VKD3D_SM4_DATA_CONTINUED */ VKD3D_DATA_CONTINUED,
|
|
/* VKD3D_SM4_DATA_UNUSED */ VKD3D_DATA_UNUSED,
|
|
};
|
|
|
|
static bool shader_is_sm_5_1(const struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
const struct vkd3d_shader_version *version = &sm4->p.program->shader_version;
|
|
|
|
return version->major >= 5 && version->minor >= 1;
|
|
}
|
|
|
|
static bool shader_sm4_read_src_param(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr,
|
|
const uint32_t *end, enum vkd3d_data_type data_type, struct vkd3d_shader_src_param *src_param);
|
|
static bool shader_sm4_read_dst_param(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr,
|
|
const uint32_t *end, enum vkd3d_data_type data_type, struct vkd3d_shader_dst_param *dst_param);
|
|
|
|
static bool shader_sm4_read_register_space(struct vkd3d_shader_sm4_parser *priv,
|
|
const uint32_t **ptr, const uint32_t *end, unsigned int *register_space)
|
|
{
|
|
*register_space = 0;
|
|
|
|
if (!shader_is_sm_5_1(priv))
|
|
return true;
|
|
|
|
if (*ptr >= end)
|
|
{
|
|
WARN("Invalid ptr %p >= end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
|
|
*register_space = *(*ptr)++;
|
|
return true;
|
|
}
|
|
|
|
static void shader_sm4_read_conditional_op(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_src_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_UINT, &ins->src[0]);
|
|
ins->flags = (opcode_token & VKD3D_SM4_CONDITIONAL_NZ) ?
|
|
VKD3D_SHADER_CONDITIONAL_OP_NZ : VKD3D_SHADER_CONDITIONAL_OP_Z;
|
|
}
|
|
|
|
static void shader_sm4_read_case_condition(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_src_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_UINT, &ins->src[0]);
|
|
if (ins->src[0].reg.type != VKD3DSPR_IMMCONST)
|
|
{
|
|
FIXME("Switch case value is not a 32-bit constant.\n");
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_CASE_VALUE,
|
|
"Switch case value is not a 32-bit immediate constant register.");
|
|
}
|
|
}
|
|
|
|
static void shader_sm4_read_shader_data(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_immediate_constant_buffer *icb;
|
|
enum vkd3d_sm4_shader_data_type type;
|
|
unsigned int icb_size;
|
|
|
|
type = (opcode_token & VKD3D_SM4_SHADER_DATA_TYPE_MASK) >> VKD3D_SM4_SHADER_DATA_TYPE_SHIFT;
|
|
if (type != VKD3D_SM4_SHADER_DATA_IMMEDIATE_CONSTANT_BUFFER)
|
|
{
|
|
FIXME("Ignoring shader data type %#x.\n", type);
|
|
ins->opcode = VKD3DSIH_NOP;
|
|
return;
|
|
}
|
|
|
|
++tokens;
|
|
icb_size = token_count - 1;
|
|
if (icb_size % 4)
|
|
{
|
|
FIXME("Unexpected immediate constant buffer size %u.\n", icb_size);
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
|
|
if (!(icb = vkd3d_malloc(offsetof(struct vkd3d_shader_immediate_constant_buffer, data[icb_size]))))
|
|
{
|
|
ERR("Failed to allocate immediate constant buffer, size %u.\n", icb_size);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_OUT_OF_MEMORY, "Out of memory.");
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
icb->register_idx = 0;
|
|
icb->data_type = VKD3D_DATA_FLOAT;
|
|
icb->component_count = VKD3D_VEC4_SIZE;
|
|
icb->element_count = icb_size / VKD3D_VEC4_SIZE;
|
|
icb->is_null = false;
|
|
memcpy(icb->data, tokens, sizeof(*tokens) * icb_size);
|
|
shader_instruction_array_add_icb(&priv->p.program->instructions, icb);
|
|
ins->declaration.icb = icb;
|
|
}
|
|
|
|
static void shader_sm4_set_descriptor_register_range(struct vkd3d_shader_sm4_parser *sm4,
|
|
const struct vkd3d_shader_register *reg, struct vkd3d_shader_register_range *range)
|
|
{
|
|
range->first = reg->idx[1].offset;
|
|
range->last = reg->idx[shader_is_sm_5_1(sm4) ? 2 : 1].offset;
|
|
if (range->last < range->first)
|
|
{
|
|
FIXME("Invalid register range [%u:%u].\n", range->first, range->last);
|
|
vkd3d_shader_parser_error(&sm4->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_RANGE,
|
|
"Last register %u must not be less than first register %u in range.", range->last, range->first);
|
|
}
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_resource(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_semantic *semantic = &ins->declaration.semantic;
|
|
enum vkd3d_sm4_resource_type resource_type;
|
|
const uint32_t *end = &tokens[token_count];
|
|
enum vkd3d_sm4_data_type data_type;
|
|
enum vkd3d_data_type reg_data_type;
|
|
uint32_t components;
|
|
unsigned int i;
|
|
|
|
resource_type = (opcode_token & VKD3D_SM4_RESOURCE_TYPE_MASK) >> VKD3D_SM4_RESOURCE_TYPE_SHIFT;
|
|
if (!resource_type || (resource_type >= ARRAY_SIZE(resource_type_table)))
|
|
{
|
|
FIXME("Unhandled resource type %#x.\n", resource_type);
|
|
semantic->resource_type = VKD3D_SHADER_RESOURCE_NONE;
|
|
}
|
|
else
|
|
{
|
|
semantic->resource_type = resource_type_table[resource_type];
|
|
}
|
|
|
|
if (semantic->resource_type == VKD3D_SHADER_RESOURCE_TEXTURE_2DMS
|
|
|| semantic->resource_type == VKD3D_SHADER_RESOURCE_TEXTURE_2DMSARRAY)
|
|
{
|
|
semantic->sample_count = (opcode_token & VKD3D_SM4_RESOURCE_SAMPLE_COUNT_MASK)
|
|
>> VKD3D_SM4_RESOURCE_SAMPLE_COUNT_SHIFT;
|
|
}
|
|
|
|
reg_data_type = VKD3D_DATA_UNUSED;
|
|
shader_sm4_read_dst_param(priv, &tokens, end, reg_data_type, &semantic->resource.reg);
|
|
shader_sm4_set_descriptor_register_range(priv, &semantic->resource.reg.reg, &semantic->resource.range);
|
|
|
|
components = *tokens++;
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; i++)
|
|
{
|
|
data_type = VKD3D_SM4_TYPE_COMPONENT(components, i);
|
|
|
|
if (!data_type || (data_type >= ARRAY_SIZE(data_type_table)))
|
|
{
|
|
FIXME("Unhandled data type %#x.\n", data_type);
|
|
semantic->resource_data_type[i] = VKD3D_DATA_FLOAT;
|
|
}
|
|
else
|
|
{
|
|
semantic->resource_data_type[i] = data_type_table[data_type];
|
|
}
|
|
}
|
|
|
|
if (opcode != VKD3D_SM4_OP_DCL_RESOURCE)
|
|
ins->flags = (opcode_token & VKD3D_SM5_UAV_FLAGS_MASK) >> VKD3D_SM5_UAV_FLAGS_SHIFT;
|
|
|
|
shader_sm4_read_register_space(priv, &tokens, end, &semantic->resource.range.space);
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_constant_buffer(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
shader_sm4_read_src_param(priv, &tokens, end, VKD3D_DATA_FLOAT, &ins->declaration.cb.src);
|
|
shader_sm4_set_descriptor_register_range(priv, &ins->declaration.cb.src.reg, &ins->declaration.cb.range);
|
|
if (opcode_token & VKD3D_SM4_INDEX_TYPE_MASK)
|
|
ins->flags |= VKD3DSI_INDEXED_DYNAMIC;
|
|
|
|
ins->declaration.cb.size = ins->declaration.cb.src.reg.idx[2].offset;
|
|
ins->declaration.cb.range.space = 0;
|
|
|
|
if (shader_is_sm_5_1(priv))
|
|
{
|
|
if (tokens >= end)
|
|
{
|
|
FIXME("Invalid ptr %p >= end %p.\n", tokens, end);
|
|
return;
|
|
}
|
|
|
|
ins->declaration.cb.size = *tokens++;
|
|
shader_sm4_read_register_space(priv, &tokens, end, &ins->declaration.cb.range.space);
|
|
}
|
|
|
|
ins->declaration.cb.size *= VKD3D_VEC4_SIZE * sizeof(float);
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_sampler(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
ins->flags = (opcode_token & VKD3D_SM4_SAMPLER_MODE_MASK) >> VKD3D_SM4_SAMPLER_MODE_SHIFT;
|
|
if (ins->flags & ~VKD3D_SM4_SAMPLER_COMPARISON)
|
|
FIXME("Unhandled sampler mode %#x.\n", ins->flags);
|
|
shader_sm4_read_src_param(priv, &tokens, end, VKD3D_DATA_UNUSED, &ins->declaration.sampler.src);
|
|
shader_sm4_set_descriptor_register_range(priv, &ins->declaration.sampler.src.reg, &ins->declaration.sampler.range);
|
|
shader_sm4_read_register_space(priv, &tokens, end, &ins->declaration.sampler.range.space);
|
|
}
|
|
|
|
static bool sm4_parser_is_in_fork_or_join_phase(const struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
return sm4->phase == VKD3DSIH_HS_FORK_PHASE || sm4->phase == VKD3DSIH_HS_JOIN_PHASE;
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_index_range(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_index_range *index_range = &ins->declaration.index_range;
|
|
struct vsir_program *program = priv->p.program;
|
|
unsigned int i, register_idx, register_count;
|
|
const struct shader_signature *signature;
|
|
enum vkd3d_shader_register_type type;
|
|
struct sm4_index_range_array *ranges;
|
|
unsigned int *io_masks;
|
|
uint32_t write_mask;
|
|
|
|
shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_OPAQUE,
|
|
&index_range->dst);
|
|
index_range->register_count = *tokens;
|
|
|
|
register_idx = index_range->dst.reg.idx[index_range->dst.reg.idx_count - 1].offset;
|
|
register_count = index_range->register_count;
|
|
write_mask = index_range->dst.write_mask;
|
|
|
|
switch ((type = index_range->dst.reg.type))
|
|
{
|
|
case VKD3DSPR_INPUT:
|
|
case VKD3DSPR_INCONTROLPOINT:
|
|
io_masks = priv->input_register_masks;
|
|
ranges = &priv->input_index_ranges;
|
|
signature = &program->input_signature;
|
|
break;
|
|
case VKD3DSPR_OUTPUT:
|
|
if (sm4_parser_is_in_fork_or_join_phase(priv))
|
|
{
|
|
io_masks = priv->patch_constant_register_masks;
|
|
ranges = &priv->patch_constant_index_ranges;
|
|
signature = &program->patch_constant_signature;
|
|
}
|
|
else
|
|
{
|
|
io_masks = priv->output_register_masks;
|
|
ranges = &priv->output_index_ranges;
|
|
signature = &program->output_signature;
|
|
}
|
|
break;
|
|
case VKD3DSPR_COLOROUT:
|
|
case VKD3DSPR_OUTCONTROLPOINT:
|
|
io_masks = priv->output_register_masks;
|
|
ranges = &priv->output_index_ranges;
|
|
signature = &program->output_signature;
|
|
break;
|
|
case VKD3DSPR_PATCHCONST:
|
|
io_masks = priv->patch_constant_register_masks;
|
|
ranges = &priv->patch_constant_index_ranges;
|
|
signature = &program->patch_constant_signature;
|
|
break;
|
|
|
|
default:
|
|
WARN("Unhandled register type %#x.\n", type);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_INDEX_RANGE_DCL,
|
|
"Invalid register type %#x for index range base %u, count %u, mask %#x.",
|
|
type, register_idx, register_count, write_mask);
|
|
return;
|
|
}
|
|
|
|
for (i = 0; i < ranges->count; ++i)
|
|
{
|
|
struct sm4_index_range r = ranges->ranges[i];
|
|
|
|
if (!(r.mask & write_mask))
|
|
continue;
|
|
/* Ranges with the same base but different lengths are not an issue. */
|
|
if (register_idx == r.index)
|
|
continue;
|
|
|
|
if ((r.index <= register_idx && register_idx - r.index < r.count)
|
|
|| (register_idx < r.index && r.index - register_idx < register_count))
|
|
{
|
|
WARN("Detected index range collision for base %u, count %u, mask %#x.\n",
|
|
register_idx, register_count, write_mask);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_INDEX_RANGE_DCL,
|
|
"Register index range base %u, count %u, mask %#x collides with a previous declaration.",
|
|
register_idx, register_count, write_mask);
|
|
return;
|
|
}
|
|
}
|
|
ranges->ranges[ranges->count].index = register_idx;
|
|
ranges->ranges[ranges->count].count = register_count;
|
|
ranges->ranges[ranges->count++].mask = write_mask;
|
|
|
|
for (i = 0; i < register_count; ++i)
|
|
{
|
|
const struct signature_element *e = vsir_signature_find_element_for_reg(signature, register_idx + i, write_mask);
|
|
/* Index ranges should not contain non-arrayed sysvals. FXC tries to forbid this but it is buggy,
|
|
* and can emit a range containing a sysval if the sysval is not actually accessed. */
|
|
if (e && e->sysval_semantic && register_count > 1 && !vsir_sysval_semantic_is_tess_factor(e->sysval_semantic)
|
|
&& !vsir_sysval_semantic_is_clip_cull(e->sysval_semantic))
|
|
{
|
|
WARN("Sysval %u included in an index range declaration.\n", e->sysval_semantic);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_INDEX_RANGE_DCL,
|
|
"Index range base %u, count %u, mask %#x contains sysval %u.",
|
|
register_idx, register_count, write_mask, e->sysval_semantic);
|
|
return;
|
|
}
|
|
if ((io_masks[register_idx + i] & write_mask) != write_mask)
|
|
{
|
|
WARN("No matching declaration for index range base %u, count %u, mask %#x.\n",
|
|
register_idx, register_count, write_mask);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_INDEX_RANGE_DCL,
|
|
"Input/output registers matching index range base %u, count %u, mask %#x were not declared.",
|
|
register_idx, register_count, write_mask);
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_output_topology(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
enum vkd3d_sm4_output_primitive_type primitive_type;
|
|
|
|
primitive_type = (opcode_token & VKD3D_SM4_PRIMITIVE_TYPE_MASK) >> VKD3D_SM4_PRIMITIVE_TYPE_SHIFT;
|
|
if (primitive_type >= ARRAY_SIZE(output_primitive_type_table))
|
|
ins->declaration.primitive_type.type = VKD3D_PT_UNDEFINED;
|
|
else
|
|
ins->declaration.primitive_type.type = output_primitive_type_table[primitive_type];
|
|
|
|
if (ins->declaration.primitive_type.type == VKD3D_PT_UNDEFINED)
|
|
FIXME("Unhandled output primitive type %#x.\n", primitive_type);
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_input_primitive(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
enum vkd3d_sm4_input_primitive_type primitive_type;
|
|
struct vsir_program *program = sm4->p.program;
|
|
|
|
primitive_type = (opcode_token & VKD3D_SM4_PRIMITIVE_TYPE_MASK) >> VKD3D_SM4_PRIMITIVE_TYPE_SHIFT;
|
|
if (VKD3D_SM5_INPUT_PT_PATCH1 <= primitive_type && primitive_type <= VKD3D_SM5_INPUT_PT_PATCH32)
|
|
{
|
|
ins->declaration.primitive_type.type = VKD3D_PT_PATCH;
|
|
ins->declaration.primitive_type.patch_vertex_count = primitive_type - VKD3D_SM5_INPUT_PT_PATCH1 + 1;
|
|
program->input_control_point_count = ins->declaration.primitive_type.patch_vertex_count;
|
|
}
|
|
else if (primitive_type >= ARRAY_SIZE(input_primitive_type_table))
|
|
{
|
|
ins->declaration.primitive_type.type = VKD3D_PT_UNDEFINED;
|
|
}
|
|
else
|
|
{
|
|
ins->declaration.primitive_type.type = input_primitive_type_table[primitive_type].vkd3d_type;
|
|
program->input_control_point_count = input_primitive_type_table[primitive_type].control_point_count;
|
|
}
|
|
|
|
if (ins->declaration.primitive_type.type == VKD3D_PT_UNDEFINED)
|
|
FIXME("Unhandled input primitive type %#x.\n", primitive_type);
|
|
}
|
|
|
|
static void shader_sm4_read_declaration_count(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
struct vsir_program *program = sm4->p.program;
|
|
|
|
ins->declaration.count = *tokens;
|
|
if (opcode == VKD3D_SM4_OP_DCL_TEMPS)
|
|
program->temp_count = max(program->temp_count, *tokens);
|
|
}
|
|
|
|
static void shader_sm4_read_declaration_dst(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT, &ins->declaration.dst);
|
|
}
|
|
|
|
static void shader_sm4_read_declaration_register_semantic(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT,
|
|
&ins->declaration.register_semantic.reg);
|
|
ins->declaration.register_semantic.sysval_semantic = *tokens;
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_input_ps(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_dst_param *dst = &ins->declaration.dst;
|
|
|
|
ins->flags = (opcode_token & VKD3D_SM4_INTERPOLATION_MODE_MASK) >> VKD3D_SM4_INTERPOLATION_MODE_SHIFT;
|
|
if (shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT, dst))
|
|
{
|
|
struct signature_element *e = vsir_signature_find_element_for_reg(
|
|
&priv->p.program->input_signature, dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
|
|
if (!e)
|
|
{
|
|
WARN("No matching signature element for input register %u with mask %#x.\n",
|
|
dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_DCL,
|
|
"No matching signature element for input register %u with mask %#x.\n",
|
|
dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
}
|
|
else
|
|
{
|
|
e->interpolation_mode = ins->flags;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_input_ps_siv(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_dst_param *dst = &ins->declaration.register_semantic.reg;
|
|
|
|
ins->flags = (opcode_token & VKD3D_SM4_INTERPOLATION_MODE_MASK) >> VKD3D_SM4_INTERPOLATION_MODE_SHIFT;
|
|
if (shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT, dst))
|
|
{
|
|
struct signature_element *e = vsir_signature_find_element_for_reg(
|
|
&priv->p.program->input_signature, dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
|
|
if (!e)
|
|
{
|
|
WARN("No matching signature element for input register %u with mask %#x.\n",
|
|
dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_DCL,
|
|
"No matching signature element for input register %u with mask %#x.\n",
|
|
dst->reg.idx[dst->reg.idx_count - 1].offset, dst->write_mask);
|
|
}
|
|
else
|
|
{
|
|
e->interpolation_mode = ins->flags;
|
|
}
|
|
}
|
|
ins->declaration.register_semantic.sysval_semantic = *tokens;
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_indexable_temp(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.indexable_temp.register_idx = *tokens++;
|
|
ins->declaration.indexable_temp.register_size = *tokens++;
|
|
ins->declaration.indexable_temp.alignment = 0;
|
|
ins->declaration.indexable_temp.data_type = VKD3D_DATA_FLOAT;
|
|
ins->declaration.indexable_temp.component_count = *tokens;
|
|
ins->declaration.indexable_temp.has_function_scope = false;
|
|
}
|
|
|
|
static void shader_sm4_read_dcl_global_flags(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.global_flags = (opcode_token & VKD3D_SM4_GLOBAL_FLAGS_MASK) >> VKD3D_SM4_GLOBAL_FLAGS_SHIFT;
|
|
}
|
|
|
|
static void shader_sm5_read_fcall(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->src[0].reg.u.fp_body_idx = *tokens++;
|
|
shader_sm4_read_src_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_OPAQUE, &ins->src[0]);
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_function_body(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.index = *tokens;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_function_table(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.index = *tokens++;
|
|
FIXME("Ignoring set of function bodies (count %u).\n", *tokens);
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_interface(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.fp.index = *tokens++;
|
|
ins->declaration.fp.body_count = *tokens++;
|
|
ins->declaration.fp.array_size = *tokens >> VKD3D_SM5_FP_ARRAY_SIZE_SHIFT;
|
|
ins->declaration.fp.table_count = *tokens++ & VKD3D_SM5_FP_TABLE_COUNT_MASK;
|
|
FIXME("Ignoring set of function tables (count %u).\n", ins->declaration.fp.table_count);
|
|
}
|
|
|
|
static void shader_sm5_read_control_point_count(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
struct vsir_program *program = sm4->p.program;
|
|
|
|
ins->declaration.count = (opcode_token & VKD3D_SM5_CONTROL_POINT_COUNT_MASK)
|
|
>> VKD3D_SM5_CONTROL_POINT_COUNT_SHIFT;
|
|
|
|
if (opcode == VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT)
|
|
program->input_control_point_count = ins->declaration.count;
|
|
else
|
|
program->output_control_point_count = ins->declaration.count;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_tessellator_domain(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.tessellator_domain = (opcode_token & VKD3D_SM5_TESSELLATOR_MASK)
|
|
>> VKD3D_SM5_TESSELLATOR_SHIFT;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_tessellator_partitioning(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.tessellator_partitioning = (opcode_token & VKD3D_SM5_TESSELLATOR_MASK)
|
|
>> VKD3D_SM5_TESSELLATOR_SHIFT;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_tessellator_output_primitive(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.tessellator_output_primitive = (opcode_token & VKD3D_SM5_TESSELLATOR_MASK)
|
|
>> VKD3D_SM5_TESSELLATOR_SHIFT;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_hs_max_tessfactor(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->declaration.max_tessellation_factor = *(float *)tokens;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_thread_group(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
struct vsir_program *program = sm4->p.program;
|
|
|
|
ins->declaration.thread_group_size.x = *tokens++;
|
|
ins->declaration.thread_group_size.y = *tokens++;
|
|
ins->declaration.thread_group_size.z = *tokens++;
|
|
program->thread_group_size = ins->declaration.thread_group_size;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_uav_raw(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_raw_resource *resource = &ins->declaration.raw_resource;
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
shader_sm4_read_dst_param(priv, &tokens, end, VKD3D_DATA_UNUSED, &resource->resource.reg);
|
|
shader_sm4_set_descriptor_register_range(priv, &resource->resource.reg.reg, &resource->resource.range);
|
|
ins->flags = (opcode_token & VKD3D_SM5_UAV_FLAGS_MASK) >> VKD3D_SM5_UAV_FLAGS_SHIFT;
|
|
shader_sm4_read_register_space(priv, &tokens, end, &resource->resource.range.space);
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_uav_structured(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_structured_resource *resource = &ins->declaration.structured_resource;
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
shader_sm4_read_dst_param(priv, &tokens, end, VKD3D_DATA_UNUSED, &resource->resource.reg);
|
|
shader_sm4_set_descriptor_register_range(priv, &resource->resource.reg.reg, &resource->resource.range);
|
|
ins->flags = (opcode_token & VKD3D_SM5_UAV_FLAGS_MASK) >> VKD3D_SM5_UAV_FLAGS_SHIFT;
|
|
resource->byte_stride = *tokens++;
|
|
if (resource->byte_stride % 4)
|
|
FIXME("Byte stride %u is not multiple of 4.\n", resource->byte_stride);
|
|
shader_sm4_read_register_space(priv, &tokens, end, &resource->resource.range.space);
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_tgsm_raw(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT, &ins->declaration.tgsm_raw.reg);
|
|
ins->declaration.tgsm_raw.byte_count = *tokens;
|
|
if (ins->declaration.tgsm_raw.byte_count % 4)
|
|
FIXME("Byte count %u is not multiple of 4.\n", ins->declaration.tgsm_raw.byte_count);
|
|
ins->declaration.tgsm_raw.zero_init = false;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_tgsm_structured(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
shader_sm4_read_dst_param(priv, &tokens, &tokens[token_count], VKD3D_DATA_FLOAT,
|
|
&ins->declaration.tgsm_structured.reg);
|
|
ins->declaration.tgsm_structured.byte_stride = *tokens++;
|
|
ins->declaration.tgsm_structured.structure_count = *tokens;
|
|
if (ins->declaration.tgsm_structured.byte_stride % 4)
|
|
FIXME("Byte stride %u is not multiple of 4.\n", ins->declaration.tgsm_structured.byte_stride);
|
|
ins->declaration.tgsm_structured.zero_init = false;
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_resource_structured(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_structured_resource *resource = &ins->declaration.structured_resource;
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
shader_sm4_read_dst_param(priv, &tokens, end, VKD3D_DATA_UNUSED, &resource->resource.reg);
|
|
shader_sm4_set_descriptor_register_range(priv, &resource->resource.reg.reg, &resource->resource.range);
|
|
resource->byte_stride = *tokens++;
|
|
if (resource->byte_stride % 4)
|
|
FIXME("Byte stride %u is not multiple of 4.\n", resource->byte_stride);
|
|
shader_sm4_read_register_space(priv, &tokens, end, &resource->resource.range.space);
|
|
}
|
|
|
|
static void shader_sm5_read_dcl_resource_raw(struct vkd3d_shader_instruction *ins, uint32_t opcode,
|
|
uint32_t opcode_token, const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
struct vkd3d_shader_raw_resource *resource = &ins->declaration.raw_resource;
|
|
const uint32_t *end = &tokens[token_count];
|
|
|
|
shader_sm4_read_dst_param(priv, &tokens, end, VKD3D_DATA_UNUSED, &resource->resource.reg);
|
|
shader_sm4_set_descriptor_register_range(priv, &resource->resource.reg.reg, &resource->resource.range);
|
|
shader_sm4_read_register_space(priv, &tokens, end, &resource->resource.range.space);
|
|
}
|
|
|
|
static void shader_sm5_read_sync(struct vkd3d_shader_instruction *ins, uint32_t opcode, uint32_t opcode_token,
|
|
const uint32_t *tokens, unsigned int token_count, struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
ins->flags = (opcode_token & VKD3D_SM5_SYNC_FLAGS_MASK) >> VKD3D_SM5_SYNC_FLAGS_SHIFT;
|
|
}
|
|
|
|
struct vkd3d_sm4_register_type_info
|
|
{
|
|
enum vkd3d_sm4_register_type sm4_type;
|
|
enum vkd3d_shader_register_type vkd3d_type;
|
|
|
|
/* Swizzle type to be used for src registers when their dimension is VKD3D_SM4_DIMENSION_VEC4. */
|
|
enum vkd3d_sm4_swizzle_type default_src_swizzle_type;
|
|
};
|
|
|
|
static const enum vkd3d_shader_register_precision register_precision_table[] =
|
|
{
|
|
/* VKD3D_SM4_REGISTER_PRECISION_DEFAULT */ VKD3D_SHADER_REGISTER_PRECISION_DEFAULT,
|
|
/* VKD3D_SM4_REGISTER_PRECISION_MIN_FLOAT_16 */ VKD3D_SHADER_REGISTER_PRECISION_MIN_FLOAT_16,
|
|
/* VKD3D_SM4_REGISTER_PRECISION_MIN_FLOAT_10 */ VKD3D_SHADER_REGISTER_PRECISION_MIN_FLOAT_10,
|
|
/* UNKNOWN */ VKD3D_SHADER_REGISTER_PRECISION_INVALID,
|
|
/* VKD3D_SM4_REGISTER_PRECISION_MIN_INT_16 */ VKD3D_SHADER_REGISTER_PRECISION_MIN_INT_16,
|
|
/* VKD3D_SM4_REGISTER_PRECISION_MIN_UINT_16 */ VKD3D_SHADER_REGISTER_PRECISION_MIN_UINT_16,
|
|
};
|
|
|
|
struct sm4_stat
|
|
{
|
|
uint32_t fields[VKD3D_STAT_COUNT];
|
|
};
|
|
|
|
struct tpf_compiler
|
|
{
|
|
/* OBJECTIVE: We want to get rid of this HLSL IR specific field. */
|
|
struct hlsl_ctx *ctx;
|
|
struct vkd3d_bytecode_buffer *buffer;
|
|
struct vkd3d_sm4_lookup_tables lookup;
|
|
struct sm4_stat *stat;
|
|
};
|
|
|
|
static void init_sm4_lookup_tables(struct vkd3d_sm4_lookup_tables *lookup)
|
|
{
|
|
unsigned int i;
|
|
|
|
/*
|
|
* d -> VKD3D_DATA_DOUBLE
|
|
* f -> VKD3D_DATA_FLOAT
|
|
* i -> VKD3D_DATA_INT
|
|
* u -> VKD3D_DATA_UINT
|
|
* O -> VKD3D_DATA_OPAQUE
|
|
* R -> VKD3D_DATA_RESOURCE
|
|
* S -> VKD3D_DATA_SAMPLER
|
|
* U -> VKD3D_DATA_UAV
|
|
*/
|
|
static const struct vkd3d_sm4_opcode_info opcode_table[] =
|
|
{
|
|
{VKD3D_SM4_OP_ADD, VKD3DSIH_ADD, "f", "ff"},
|
|
{VKD3D_SM4_OP_AND, VKD3DSIH_AND, "u", "uu"},
|
|
{VKD3D_SM4_OP_BREAK, VKD3DSIH_BREAK, "", ""},
|
|
{VKD3D_SM4_OP_BREAKC, VKD3DSIH_BREAKP, "", "u",
|
|
shader_sm4_read_conditional_op},
|
|
{VKD3D_SM4_OP_CASE, VKD3DSIH_CASE, "", "u",
|
|
shader_sm4_read_case_condition},
|
|
{VKD3D_SM4_OP_CONTINUE, VKD3DSIH_CONTINUE, "", ""},
|
|
{VKD3D_SM4_OP_CONTINUEC, VKD3DSIH_CONTINUEP, "", "u",
|
|
shader_sm4_read_conditional_op},
|
|
{VKD3D_SM4_OP_CUT, VKD3DSIH_CUT, "", ""},
|
|
{VKD3D_SM4_OP_DEFAULT, VKD3DSIH_DEFAULT, "", ""},
|
|
{VKD3D_SM4_OP_DERIV_RTX, VKD3DSIH_DSX, "f", "f"},
|
|
{VKD3D_SM4_OP_DERIV_RTY, VKD3DSIH_DSY, "f", "f"},
|
|
{VKD3D_SM4_OP_DISCARD, VKD3DSIH_DISCARD, "", "u",
|
|
shader_sm4_read_conditional_op},
|
|
{VKD3D_SM4_OP_DIV, VKD3DSIH_DIV, "f", "ff"},
|
|
{VKD3D_SM4_OP_DP2, VKD3DSIH_DP2, "f", "ff"},
|
|
{VKD3D_SM4_OP_DP3, VKD3DSIH_DP3, "f", "ff"},
|
|
{VKD3D_SM4_OP_DP4, VKD3DSIH_DP4, "f", "ff"},
|
|
{VKD3D_SM4_OP_ELSE, VKD3DSIH_ELSE, "", ""},
|
|
{VKD3D_SM4_OP_EMIT, VKD3DSIH_EMIT, "", ""},
|
|
{VKD3D_SM4_OP_ENDIF, VKD3DSIH_ENDIF, "", ""},
|
|
{VKD3D_SM4_OP_ENDLOOP, VKD3DSIH_ENDLOOP, "", ""},
|
|
{VKD3D_SM4_OP_ENDSWITCH, VKD3DSIH_ENDSWITCH, "", ""},
|
|
{VKD3D_SM4_OP_EQ, VKD3DSIH_EQO, "u", "ff"},
|
|
{VKD3D_SM4_OP_EXP, VKD3DSIH_EXP, "f", "f"},
|
|
{VKD3D_SM4_OP_FRC, VKD3DSIH_FRC, "f", "f"},
|
|
{VKD3D_SM4_OP_FTOI, VKD3DSIH_FTOI, "i", "f"},
|
|
{VKD3D_SM4_OP_FTOU, VKD3DSIH_FTOU, "u", "f"},
|
|
{VKD3D_SM4_OP_GE, VKD3DSIH_GEO, "u", "ff"},
|
|
{VKD3D_SM4_OP_IADD, VKD3DSIH_IADD, "i", "ii"},
|
|
{VKD3D_SM4_OP_IF, VKD3DSIH_IF, "", "u",
|
|
shader_sm4_read_conditional_op},
|
|
{VKD3D_SM4_OP_IEQ, VKD3DSIH_IEQ, "u", "ii"},
|
|
{VKD3D_SM4_OP_IGE, VKD3DSIH_IGE, "u", "ii"},
|
|
{VKD3D_SM4_OP_ILT, VKD3DSIH_ILT, "u", "ii"},
|
|
{VKD3D_SM4_OP_IMAD, VKD3DSIH_IMAD, "i", "iii"},
|
|
{VKD3D_SM4_OP_IMAX, VKD3DSIH_IMAX, "i", "ii"},
|
|
{VKD3D_SM4_OP_IMIN, VKD3DSIH_IMIN, "i", "ii"},
|
|
{VKD3D_SM4_OP_IMUL, VKD3DSIH_IMUL, "ii", "ii"},
|
|
{VKD3D_SM4_OP_INE, VKD3DSIH_INE, "u", "ii"},
|
|
{VKD3D_SM4_OP_INEG, VKD3DSIH_INEG, "i", "i"},
|
|
{VKD3D_SM4_OP_ISHL, VKD3DSIH_ISHL, "i", "ii"},
|
|
{VKD3D_SM4_OP_ISHR, VKD3DSIH_ISHR, "i", "ii"},
|
|
{VKD3D_SM4_OP_ITOF, VKD3DSIH_ITOF, "f", "i"},
|
|
{VKD3D_SM4_OP_LABEL, VKD3DSIH_LABEL, "", "O"},
|
|
{VKD3D_SM4_OP_LD, VKD3DSIH_LD, "u", "i*"},
|
|
{VKD3D_SM4_OP_LD2DMS, VKD3DSIH_LD2DMS, "u", "i*i"},
|
|
{VKD3D_SM4_OP_LOG, VKD3DSIH_LOG, "f", "f"},
|
|
{VKD3D_SM4_OP_LOOP, VKD3DSIH_LOOP, "", ""},
|
|
{VKD3D_SM4_OP_LT, VKD3DSIH_LTO, "u", "ff"},
|
|
{VKD3D_SM4_OP_MAD, VKD3DSIH_MAD, "f", "fff"},
|
|
{VKD3D_SM4_OP_MIN, VKD3DSIH_MIN, "f", "ff"},
|
|
{VKD3D_SM4_OP_MAX, VKD3DSIH_MAX, "f", "ff"},
|
|
{VKD3D_SM4_OP_SHADER_DATA, VKD3DSIH_DCL_IMMEDIATE_CONSTANT_BUFFER, "", "",
|
|
shader_sm4_read_shader_data},
|
|
{VKD3D_SM4_OP_MOV, VKD3DSIH_MOV, "f", "f"},
|
|
{VKD3D_SM4_OP_MOVC, VKD3DSIH_MOVC, "f", "uff"},
|
|
{VKD3D_SM4_OP_MUL, VKD3DSIH_MUL, "f", "ff"},
|
|
{VKD3D_SM4_OP_NE, VKD3DSIH_NEU, "u", "ff"},
|
|
{VKD3D_SM4_OP_NOP, VKD3DSIH_NOP, "", ""},
|
|
{VKD3D_SM4_OP_NOT, VKD3DSIH_NOT, "u", "u"},
|
|
{VKD3D_SM4_OP_OR, VKD3DSIH_OR, "u", "uu"},
|
|
{VKD3D_SM4_OP_RESINFO, VKD3DSIH_RESINFO, "f", "i*"},
|
|
{VKD3D_SM4_OP_RET, VKD3DSIH_RET, "", ""},
|
|
{VKD3D_SM4_OP_RETC, VKD3DSIH_RETP, "", "u",
|
|
shader_sm4_read_conditional_op},
|
|
{VKD3D_SM4_OP_ROUND_NE, VKD3DSIH_ROUND_NE, "f", "f"},
|
|
{VKD3D_SM4_OP_ROUND_NI, VKD3DSIH_ROUND_NI, "f", "f"},
|
|
{VKD3D_SM4_OP_ROUND_PI, VKD3DSIH_ROUND_PI, "f", "f"},
|
|
{VKD3D_SM4_OP_ROUND_Z, VKD3DSIH_ROUND_Z, "f", "f"},
|
|
{VKD3D_SM4_OP_RSQ, VKD3DSIH_RSQ, "f", "f"},
|
|
{VKD3D_SM4_OP_SAMPLE, VKD3DSIH_SAMPLE, "u", "f**"},
|
|
{VKD3D_SM4_OP_SAMPLE_C, VKD3DSIH_SAMPLE_C, "f", "f**f"},
|
|
{VKD3D_SM4_OP_SAMPLE_C_LZ, VKD3DSIH_SAMPLE_C_LZ, "f", "f**f"},
|
|
{VKD3D_SM4_OP_SAMPLE_LOD, VKD3DSIH_SAMPLE_LOD, "u", "f**f"},
|
|
{VKD3D_SM4_OP_SAMPLE_GRAD, VKD3DSIH_SAMPLE_GRAD, "u", "f**ff"},
|
|
{VKD3D_SM4_OP_SAMPLE_B, VKD3DSIH_SAMPLE_B, "u", "f**f"},
|
|
{VKD3D_SM4_OP_SQRT, VKD3DSIH_SQRT, "f", "f"},
|
|
{VKD3D_SM4_OP_SWITCH, VKD3DSIH_SWITCH, "", "i"},
|
|
{VKD3D_SM4_OP_SINCOS, VKD3DSIH_SINCOS, "ff", "f"},
|
|
{VKD3D_SM4_OP_UDIV, VKD3DSIH_UDIV, "uu", "uu"},
|
|
{VKD3D_SM4_OP_ULT, VKD3DSIH_ULT, "u", "uu"},
|
|
{VKD3D_SM4_OP_UGE, VKD3DSIH_UGE, "u", "uu"},
|
|
{VKD3D_SM4_OP_UMUL, VKD3DSIH_UMUL, "uu", "uu"},
|
|
{VKD3D_SM4_OP_UMAX, VKD3DSIH_UMAX, "u", "uu"},
|
|
{VKD3D_SM4_OP_UMIN, VKD3DSIH_UMIN, "u", "uu"},
|
|
{VKD3D_SM4_OP_USHR, VKD3DSIH_USHR, "u", "uu"},
|
|
{VKD3D_SM4_OP_UTOF, VKD3DSIH_UTOF, "f", "u"},
|
|
{VKD3D_SM4_OP_XOR, VKD3DSIH_XOR, "u", "uu"},
|
|
{VKD3D_SM4_OP_DCL_RESOURCE, VKD3DSIH_DCL, "", "",
|
|
shader_sm4_read_dcl_resource},
|
|
{VKD3D_SM4_OP_DCL_CONSTANT_BUFFER, VKD3DSIH_DCL_CONSTANT_BUFFER, "", "",
|
|
shader_sm4_read_dcl_constant_buffer},
|
|
{VKD3D_SM4_OP_DCL_SAMPLER, VKD3DSIH_DCL_SAMPLER, "", "",
|
|
shader_sm4_read_dcl_sampler},
|
|
{VKD3D_SM4_OP_DCL_INDEX_RANGE, VKD3DSIH_DCL_INDEX_RANGE, "", "",
|
|
shader_sm4_read_dcl_index_range},
|
|
{VKD3D_SM4_OP_DCL_OUTPUT_TOPOLOGY, VKD3DSIH_DCL_OUTPUT_TOPOLOGY, "", "",
|
|
shader_sm4_read_dcl_output_topology},
|
|
{VKD3D_SM4_OP_DCL_INPUT_PRIMITIVE, VKD3DSIH_DCL_INPUT_PRIMITIVE, "", "",
|
|
shader_sm4_read_dcl_input_primitive},
|
|
{VKD3D_SM4_OP_DCL_VERTICES_OUT, VKD3DSIH_DCL_VERTICES_OUT, "", "",
|
|
shader_sm4_read_declaration_count},
|
|
{VKD3D_SM4_OP_DCL_INPUT, VKD3DSIH_DCL_INPUT, "", "",
|
|
shader_sm4_read_declaration_dst},
|
|
{VKD3D_SM4_OP_DCL_INPUT_SGV, VKD3DSIH_DCL_INPUT_SGV, "", "",
|
|
shader_sm4_read_declaration_register_semantic},
|
|
{VKD3D_SM4_OP_DCL_INPUT_SIV, VKD3DSIH_DCL_INPUT_SIV, "", "",
|
|
shader_sm4_read_declaration_register_semantic},
|
|
{VKD3D_SM4_OP_DCL_INPUT_PS, VKD3DSIH_DCL_INPUT_PS, "", "",
|
|
shader_sm4_read_dcl_input_ps},
|
|
{VKD3D_SM4_OP_DCL_INPUT_PS_SGV, VKD3DSIH_DCL_INPUT_PS_SGV, "", "",
|
|
shader_sm4_read_declaration_register_semantic},
|
|
{VKD3D_SM4_OP_DCL_INPUT_PS_SIV, VKD3DSIH_DCL_INPUT_PS_SIV, "", "",
|
|
shader_sm4_read_dcl_input_ps_siv},
|
|
{VKD3D_SM4_OP_DCL_OUTPUT, VKD3DSIH_DCL_OUTPUT, "", "",
|
|
shader_sm4_read_declaration_dst},
|
|
{VKD3D_SM4_OP_DCL_OUTPUT_SIV, VKD3DSIH_DCL_OUTPUT_SIV, "", "",
|
|
shader_sm4_read_declaration_register_semantic},
|
|
{VKD3D_SM4_OP_DCL_TEMPS, VKD3DSIH_DCL_TEMPS, "", "",
|
|
shader_sm4_read_declaration_count},
|
|
{VKD3D_SM4_OP_DCL_INDEXABLE_TEMP, VKD3DSIH_DCL_INDEXABLE_TEMP, "", "",
|
|
shader_sm4_read_dcl_indexable_temp},
|
|
{VKD3D_SM4_OP_DCL_GLOBAL_FLAGS, VKD3DSIH_DCL_GLOBAL_FLAGS, "", "",
|
|
shader_sm4_read_dcl_global_flags},
|
|
{VKD3D_SM4_OP_LOD, VKD3DSIH_LOD, "f", "f**"},
|
|
{VKD3D_SM4_OP_GATHER4, VKD3DSIH_GATHER4, "u", "f**"},
|
|
{VKD3D_SM4_OP_SAMPLE_POS, VKD3DSIH_SAMPLE_POS, "f", "*u"},
|
|
{VKD3D_SM4_OP_SAMPLE_INFO, VKD3DSIH_SAMPLE_INFO, "f", "*"},
|
|
{VKD3D_SM5_OP_HS_DECLS, VKD3DSIH_HS_DECLS, "", ""},
|
|
{VKD3D_SM5_OP_HS_CONTROL_POINT_PHASE, VKD3DSIH_HS_CONTROL_POINT_PHASE, "", ""},
|
|
{VKD3D_SM5_OP_HS_FORK_PHASE, VKD3DSIH_HS_FORK_PHASE, "", ""},
|
|
{VKD3D_SM5_OP_HS_JOIN_PHASE, VKD3DSIH_HS_JOIN_PHASE, "", ""},
|
|
{VKD3D_SM5_OP_EMIT_STREAM, VKD3DSIH_EMIT_STREAM, "", "f"},
|
|
{VKD3D_SM5_OP_CUT_STREAM, VKD3DSIH_CUT_STREAM, "", "f"},
|
|
{VKD3D_SM5_OP_FCALL, VKD3DSIH_FCALL, "", "O",
|
|
shader_sm5_read_fcall},
|
|
{VKD3D_SM5_OP_BUFINFO, VKD3DSIH_BUFINFO, "i", "*"},
|
|
{VKD3D_SM5_OP_DERIV_RTX_COARSE, VKD3DSIH_DSX_COARSE, "f", "f"},
|
|
{VKD3D_SM5_OP_DERIV_RTX_FINE, VKD3DSIH_DSX_FINE, "f", "f"},
|
|
{VKD3D_SM5_OP_DERIV_RTY_COARSE, VKD3DSIH_DSY_COARSE, "f", "f"},
|
|
{VKD3D_SM5_OP_DERIV_RTY_FINE, VKD3DSIH_DSY_FINE, "f", "f"},
|
|
{VKD3D_SM5_OP_GATHER4_C, VKD3DSIH_GATHER4_C, "f", "f**f"},
|
|
{VKD3D_SM5_OP_GATHER4_PO, VKD3DSIH_GATHER4_PO, "f", "fi**"},
|
|
{VKD3D_SM5_OP_GATHER4_PO_C, VKD3DSIH_GATHER4_PO_C, "f", "fi**f"},
|
|
{VKD3D_SM5_OP_RCP, VKD3DSIH_RCP, "f", "f"},
|
|
{VKD3D_SM5_OP_F32TOF16, VKD3DSIH_F32TOF16, "u", "f"},
|
|
{VKD3D_SM5_OP_F16TOF32, VKD3DSIH_F16TOF32, "f", "u"},
|
|
{VKD3D_SM5_OP_COUNTBITS, VKD3DSIH_COUNTBITS, "u", "u"},
|
|
{VKD3D_SM5_OP_FIRSTBIT_HI, VKD3DSIH_FIRSTBIT_HI, "u", "u"},
|
|
{VKD3D_SM5_OP_FIRSTBIT_LO, VKD3DSIH_FIRSTBIT_LO, "u", "u"},
|
|
{VKD3D_SM5_OP_FIRSTBIT_SHI, VKD3DSIH_FIRSTBIT_SHI, "u", "i"},
|
|
{VKD3D_SM5_OP_UBFE, VKD3DSIH_UBFE, "u", "iiu"},
|
|
{VKD3D_SM5_OP_IBFE, VKD3DSIH_IBFE, "i", "iii"},
|
|
{VKD3D_SM5_OP_BFI, VKD3DSIH_BFI, "u", "iiuu"},
|
|
{VKD3D_SM5_OP_BFREV, VKD3DSIH_BFREV, "u", "u"},
|
|
{VKD3D_SM5_OP_SWAPC, VKD3DSIH_SWAPC, "ff", "uff"},
|
|
{VKD3D_SM5_OP_DCL_STREAM, VKD3DSIH_DCL_STREAM, "", "O"},
|
|
{VKD3D_SM5_OP_DCL_FUNCTION_BODY, VKD3DSIH_DCL_FUNCTION_BODY, "", "",
|
|
shader_sm5_read_dcl_function_body},
|
|
{VKD3D_SM5_OP_DCL_FUNCTION_TABLE, VKD3DSIH_DCL_FUNCTION_TABLE, "", "",
|
|
shader_sm5_read_dcl_function_table},
|
|
{VKD3D_SM5_OP_DCL_INTERFACE, VKD3DSIH_DCL_INTERFACE, "", "",
|
|
shader_sm5_read_dcl_interface},
|
|
{VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT, VKD3DSIH_DCL_INPUT_CONTROL_POINT_COUNT, "", "",
|
|
shader_sm5_read_control_point_count},
|
|
{VKD3D_SM5_OP_DCL_OUTPUT_CONTROL_POINT_COUNT, VKD3DSIH_DCL_OUTPUT_CONTROL_POINT_COUNT, "", "",
|
|
shader_sm5_read_control_point_count},
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_DOMAIN, VKD3DSIH_DCL_TESSELLATOR_DOMAIN, "", "",
|
|
shader_sm5_read_dcl_tessellator_domain},
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_PARTITIONING, VKD3DSIH_DCL_TESSELLATOR_PARTITIONING, "", "",
|
|
shader_sm5_read_dcl_tessellator_partitioning},
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_OUTPUT_PRIMITIVE, VKD3DSIH_DCL_TESSELLATOR_OUTPUT_PRIMITIVE, "", "",
|
|
shader_sm5_read_dcl_tessellator_output_primitive},
|
|
{VKD3D_SM5_OP_DCL_HS_MAX_TESSFACTOR, VKD3DSIH_DCL_HS_MAX_TESSFACTOR, "", "",
|
|
shader_sm5_read_dcl_hs_max_tessfactor},
|
|
{VKD3D_SM5_OP_DCL_HS_FORK_PHASE_INSTANCE_COUNT, VKD3DSIH_DCL_HS_FORK_PHASE_INSTANCE_COUNT, "", "",
|
|
shader_sm4_read_declaration_count},
|
|
{VKD3D_SM5_OP_DCL_HS_JOIN_PHASE_INSTANCE_COUNT, VKD3DSIH_DCL_HS_JOIN_PHASE_INSTANCE_COUNT, "", "",
|
|
shader_sm4_read_declaration_count},
|
|
{VKD3D_SM5_OP_DCL_THREAD_GROUP, VKD3DSIH_DCL_THREAD_GROUP, "", "",
|
|
shader_sm5_read_dcl_thread_group},
|
|
{VKD3D_SM5_OP_DCL_UAV_TYPED, VKD3DSIH_DCL_UAV_TYPED, "", "",
|
|
shader_sm4_read_dcl_resource},
|
|
{VKD3D_SM5_OP_DCL_UAV_RAW, VKD3DSIH_DCL_UAV_RAW, "", "",
|
|
shader_sm5_read_dcl_uav_raw},
|
|
{VKD3D_SM5_OP_DCL_UAV_STRUCTURED, VKD3DSIH_DCL_UAV_STRUCTURED, "", "",
|
|
shader_sm5_read_dcl_uav_structured},
|
|
{VKD3D_SM5_OP_DCL_TGSM_RAW, VKD3DSIH_DCL_TGSM_RAW, "", "",
|
|
shader_sm5_read_dcl_tgsm_raw},
|
|
{VKD3D_SM5_OP_DCL_TGSM_STRUCTURED, VKD3DSIH_DCL_TGSM_STRUCTURED, "", "",
|
|
shader_sm5_read_dcl_tgsm_structured},
|
|
{VKD3D_SM5_OP_DCL_RESOURCE_RAW, VKD3DSIH_DCL_RESOURCE_RAW, "", "",
|
|
shader_sm5_read_dcl_resource_raw},
|
|
{VKD3D_SM5_OP_DCL_RESOURCE_STRUCTURED, VKD3DSIH_DCL_RESOURCE_STRUCTURED, "", "",
|
|
shader_sm5_read_dcl_resource_structured},
|
|
{VKD3D_SM5_OP_LD_UAV_TYPED, VKD3DSIH_LD_UAV_TYPED, "u", "i*"},
|
|
{VKD3D_SM5_OP_STORE_UAV_TYPED, VKD3DSIH_STORE_UAV_TYPED, "*", "iu"},
|
|
{VKD3D_SM5_OP_LD_RAW, VKD3DSIH_LD_RAW, "u", "i*"},
|
|
{VKD3D_SM5_OP_STORE_RAW, VKD3DSIH_STORE_RAW, "*", "uu"},
|
|
{VKD3D_SM5_OP_LD_STRUCTURED, VKD3DSIH_LD_STRUCTURED, "u", "ii*"},
|
|
{VKD3D_SM5_OP_STORE_STRUCTURED, VKD3DSIH_STORE_STRUCTURED, "*", "iiu"},
|
|
{VKD3D_SM5_OP_ATOMIC_AND, VKD3DSIH_ATOMIC_AND, "*", "iu"},
|
|
{VKD3D_SM5_OP_ATOMIC_OR, VKD3DSIH_ATOMIC_OR, "*", "iu"},
|
|
{VKD3D_SM5_OP_ATOMIC_XOR, VKD3DSIH_ATOMIC_XOR, "*", "iu"},
|
|
{VKD3D_SM5_OP_ATOMIC_CMP_STORE, VKD3DSIH_ATOMIC_CMP_STORE, "*", "iuu"},
|
|
{VKD3D_SM5_OP_ATOMIC_IADD, VKD3DSIH_ATOMIC_IADD, "*", "ii"},
|
|
{VKD3D_SM5_OP_ATOMIC_IMAX, VKD3DSIH_ATOMIC_IMAX, "*", "ii"},
|
|
{VKD3D_SM5_OP_ATOMIC_IMIN, VKD3DSIH_ATOMIC_IMIN, "*", "ii"},
|
|
{VKD3D_SM5_OP_ATOMIC_UMAX, VKD3DSIH_ATOMIC_UMAX, "*", "iu"},
|
|
{VKD3D_SM5_OP_ATOMIC_UMIN, VKD3DSIH_ATOMIC_UMIN, "*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_ALLOC, VKD3DSIH_IMM_ATOMIC_ALLOC, "u", "*"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_CONSUME, VKD3DSIH_IMM_ATOMIC_CONSUME, "u", "*"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IADD, VKD3DSIH_IMM_ATOMIC_IADD, "u*", "ii"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_AND, VKD3DSIH_IMM_ATOMIC_AND, "u*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_OR, VKD3DSIH_IMM_ATOMIC_OR, "u*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_XOR, VKD3DSIH_IMM_ATOMIC_XOR, "u*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_EXCH, VKD3DSIH_IMM_ATOMIC_EXCH, "u*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_CMP_EXCH, VKD3DSIH_IMM_ATOMIC_CMP_EXCH, "u*", "iuu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IMAX, VKD3DSIH_IMM_ATOMIC_IMAX, "i*", "ii"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IMIN, VKD3DSIH_IMM_ATOMIC_IMIN, "i*", "ii"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_UMAX, VKD3DSIH_IMM_ATOMIC_UMAX, "u*", "iu"},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_UMIN, VKD3DSIH_IMM_ATOMIC_UMIN, "u*", "iu"},
|
|
{VKD3D_SM5_OP_SYNC, VKD3DSIH_SYNC, "", "",
|
|
shader_sm5_read_sync},
|
|
{VKD3D_SM5_OP_DADD, VKD3DSIH_DADD, "d", "dd"},
|
|
{VKD3D_SM5_OP_DMAX, VKD3DSIH_DMAX, "d", "dd"},
|
|
{VKD3D_SM5_OP_DMIN, VKD3DSIH_DMIN, "d", "dd"},
|
|
{VKD3D_SM5_OP_DMUL, VKD3DSIH_DMUL, "d", "dd"},
|
|
{VKD3D_SM5_OP_DEQ, VKD3DSIH_DEQO, "u", "dd"},
|
|
{VKD3D_SM5_OP_DGE, VKD3DSIH_DGEO, "u", "dd"},
|
|
{VKD3D_SM5_OP_DLT, VKD3DSIH_DLT, "u", "dd"},
|
|
{VKD3D_SM5_OP_DNE, VKD3DSIH_DNE, "u", "dd"},
|
|
{VKD3D_SM5_OP_DMOV, VKD3DSIH_DMOV, "d", "d"},
|
|
{VKD3D_SM5_OP_DMOVC, VKD3DSIH_DMOVC, "d", "udd"},
|
|
{VKD3D_SM5_OP_DTOF, VKD3DSIH_DTOF, "f", "d"},
|
|
{VKD3D_SM5_OP_FTOD, VKD3DSIH_FTOD, "d", "f"},
|
|
{VKD3D_SM5_OP_EVAL_SAMPLE_INDEX, VKD3DSIH_EVAL_SAMPLE_INDEX, "f", "fi"},
|
|
{VKD3D_SM5_OP_EVAL_CENTROID, VKD3DSIH_EVAL_CENTROID, "f", "f"},
|
|
{VKD3D_SM5_OP_DCL_GS_INSTANCES, VKD3DSIH_DCL_GS_INSTANCES, "", "",
|
|
shader_sm4_read_declaration_count},
|
|
{VKD3D_SM5_OP_DDIV, VKD3DSIH_DDIV, "d", "dd"},
|
|
{VKD3D_SM5_OP_DFMA, VKD3DSIH_DFMA, "d", "ddd"},
|
|
{VKD3D_SM5_OP_DRCP, VKD3DSIH_DRCP, "d", "d"},
|
|
{VKD3D_SM5_OP_MSAD, VKD3DSIH_MSAD, "u", "uuu"},
|
|
{VKD3D_SM5_OP_DTOI, VKD3DSIH_DTOI, "i", "d"},
|
|
{VKD3D_SM5_OP_DTOU, VKD3DSIH_DTOU, "u", "d"},
|
|
{VKD3D_SM5_OP_ITOD, VKD3DSIH_ITOD, "d", "i"},
|
|
{VKD3D_SM5_OP_UTOD, VKD3DSIH_UTOD, "d", "u"},
|
|
{VKD3D_SM5_OP_GATHER4_S, VKD3DSIH_GATHER4_S, "uu", "f**"},
|
|
{VKD3D_SM5_OP_GATHER4_C_S, VKD3DSIH_GATHER4_C_S, "fu", "f**f"},
|
|
{VKD3D_SM5_OP_GATHER4_PO_S, VKD3DSIH_GATHER4_PO_S, "fu", "fi**"},
|
|
{VKD3D_SM5_OP_GATHER4_PO_C_S, VKD3DSIH_GATHER4_PO_C_S, "fu", "fi**f"},
|
|
{VKD3D_SM5_OP_LD_S, VKD3DSIH_LD_S, "uu", "i*"},
|
|
{VKD3D_SM5_OP_LD2DMS_S, VKD3DSIH_LD2DMS_S, "uu", "i*i"},
|
|
{VKD3D_SM5_OP_LD_UAV_TYPED_S, VKD3DSIH_LD_UAV_TYPED_S, "uu", "iU"},
|
|
{VKD3D_SM5_OP_LD_RAW_S, VKD3DSIH_LD_RAW_S, "uu", "iU"},
|
|
{VKD3D_SM5_OP_LD_STRUCTURED_S, VKD3DSIH_LD_STRUCTURED_S, "uu", "ii*"},
|
|
{VKD3D_SM5_OP_SAMPLE_LOD_S, VKD3DSIH_SAMPLE_LOD_S, "uu", "f**f"},
|
|
{VKD3D_SM5_OP_SAMPLE_C_LZ_S, VKD3DSIH_SAMPLE_C_LZ_S, "fu", "f**f"},
|
|
{VKD3D_SM5_OP_SAMPLE_CL_S, VKD3DSIH_SAMPLE_CL_S, "uu", "f**f"},
|
|
{VKD3D_SM5_OP_SAMPLE_B_CL_S, VKD3DSIH_SAMPLE_B_CL_S, "uu", "f**ff"},
|
|
{VKD3D_SM5_OP_SAMPLE_GRAD_CL_S, VKD3DSIH_SAMPLE_GRAD_CL_S, "uu", "f**fff"},
|
|
{VKD3D_SM5_OP_SAMPLE_C_CL_S, VKD3DSIH_SAMPLE_C_CL_S, "fu", "f**ff"},
|
|
{VKD3D_SM5_OP_CHECK_ACCESS_FULLY_MAPPED, VKD3DSIH_CHECK_ACCESS_FULLY_MAPPED, "u", "u"},
|
|
};
|
|
|
|
static const struct vkd3d_sm4_register_type_info register_type_table[] =
|
|
{
|
|
{VKD3D_SM4_RT_TEMP, VKD3DSPR_TEMP, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_INPUT, VKD3DSPR_INPUT, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_OUTPUT, VKD3DSPR_OUTPUT, VKD3D_SM4_SWIZZLE_INVALID},
|
|
{VKD3D_SM4_RT_INDEXABLE_TEMP, VKD3DSPR_IDXTEMP, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_IMMCONST, VKD3DSPR_IMMCONST, VKD3D_SM4_SWIZZLE_NONE},
|
|
{VKD3D_SM4_RT_IMMCONST64, VKD3DSPR_IMMCONST64, VKD3D_SM4_SWIZZLE_NONE},
|
|
{VKD3D_SM4_RT_SAMPLER, VKD3DSPR_SAMPLER, VKD3D_SM4_SWIZZLE_SCALAR},
|
|
{VKD3D_SM4_RT_RESOURCE, VKD3DSPR_RESOURCE, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_CONSTBUFFER, VKD3DSPR_CONSTBUFFER, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_IMMCONSTBUFFER, VKD3DSPR_IMMCONSTBUFFER, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_PRIMID, VKD3DSPR_PRIMID, VKD3D_SM4_SWIZZLE_NONE},
|
|
{VKD3D_SM4_RT_DEPTHOUT, VKD3DSPR_DEPTHOUT, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_NULL, VKD3DSPR_NULL, VKD3D_SM4_SWIZZLE_INVALID},
|
|
{VKD3D_SM4_RT_RASTERIZER, VKD3DSPR_RASTERIZER, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM4_RT_OMASK, VKD3DSPR_SAMPLEMASK, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_STREAM, VKD3DSPR_STREAM, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_FUNCTION_BODY, VKD3DSPR_FUNCTIONBODY, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_FUNCTION_POINTER, VKD3DSPR_FUNCTIONPOINTER, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_OUTPUT_CONTROL_POINT_ID, VKD3DSPR_OUTPOINTID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_FORK_INSTANCE_ID, VKD3DSPR_FORKINSTID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_JOIN_INSTANCE_ID, VKD3DSPR_JOININSTID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_INPUT_CONTROL_POINT, VKD3DSPR_INCONTROLPOINT, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_OUTPUT_CONTROL_POINT, VKD3DSPR_OUTCONTROLPOINT, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_PATCH_CONSTANT_DATA, VKD3DSPR_PATCHCONST, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_DOMAIN_LOCATION, VKD3DSPR_TESSCOORD, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_UAV, VKD3DSPR_UAV, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_SHARED_MEMORY, VKD3DSPR_GROUPSHAREDMEM, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_THREAD_ID, VKD3DSPR_THREADID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_THREAD_GROUP_ID, VKD3DSPR_THREADGROUPID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_LOCAL_THREAD_ID, VKD3DSPR_LOCALTHREADID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_COVERAGE, VKD3DSPR_COVERAGE, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_LOCAL_THREAD_INDEX, VKD3DSPR_LOCALTHREADINDEX,VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_GS_INSTANCE_ID, VKD3DSPR_GSINSTID, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_DEPTHOUT_GREATER_EQUAL, VKD3DSPR_DEPTHOUTGE, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_DEPTHOUT_LESS_EQUAL, VKD3DSPR_DEPTHOUTLE, VKD3D_SM4_SWIZZLE_VEC4},
|
|
{VKD3D_SM5_RT_OUTPUT_STENCIL_REF, VKD3DSPR_OUTSTENCILREF, VKD3D_SM4_SWIZZLE_VEC4},
|
|
};
|
|
|
|
static const struct vkd3d_sm4_stat_field_info stat_field_table[] =
|
|
{
|
|
{VKD3D_SM4_OP_MOV, VKD3D_STAT_MOV},
|
|
{VKD3D_SM4_OP_MOVC, VKD3D_STAT_MOVC},
|
|
{VKD3D_SM5_OP_DMOV, VKD3D_STAT_MOV},
|
|
{VKD3D_SM5_OP_DMOVC, VKD3D_STAT_MOVC},
|
|
|
|
{VKD3D_SM4_OP_ITOF, VKD3D_STAT_CONV},
|
|
{VKD3D_SM4_OP_FTOI, VKD3D_STAT_CONV},
|
|
{VKD3D_SM4_OP_FTOU, VKD3D_STAT_CONV},
|
|
{VKD3D_SM4_OP_UTOF, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_DTOU, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_UTOD, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_DTOF, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_FTOD, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_DTOI, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_ITOD, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_F32TOF16, VKD3D_STAT_CONV},
|
|
{VKD3D_SM5_OP_F16TOF32, VKD3D_STAT_CONV},
|
|
|
|
{VKD3D_SM4_OP_ADD, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_DIV, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_DP2, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_DP3, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_DP4, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_EQ, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_EXP, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_FRC, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_GE, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_LT, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_MAD, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_MIN, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_MAX, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_MUL, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_NE, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_ROUND_NE, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_ROUND_NI, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_ROUND_PI, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_ROUND_Z, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_RSQ, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_SQRT, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM4_OP_SINCOS, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_RCP, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DADD, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DMAX, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DMIN, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DMUL, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DEQ, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DGE, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DLT, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DNE, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DDIV, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DFMA, VKD3D_STAT_FLOAT},
|
|
{VKD3D_SM5_OP_DRCP, VKD3D_STAT_FLOAT},
|
|
|
|
{VKD3D_SM4_OP_IADD, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IEQ, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IGE, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_ILT, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IMAD, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IMAX, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IMIN, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_IMUL, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_INE, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_INEG, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_ISHL, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_ISHR, VKD3D_STAT_INT},
|
|
{VKD3D_SM4_OP_ITOF, VKD3D_STAT_INT},
|
|
|
|
{VKD3D_SM4_OP_UDIV, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_ULT, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_UGE, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_UMUL, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_UMAX, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_UMIN, VKD3D_STAT_UINT},
|
|
{VKD3D_SM4_OP_USHR, VKD3D_STAT_UINT},
|
|
|
|
{VKD3D_SM4_OP_EMIT, VKD3D_STAT_EMIT},
|
|
{VKD3D_SM4_OP_CUT, VKD3D_STAT_CUT},
|
|
{VKD3D_SM5_OP_EMIT_STREAM, VKD3D_STAT_EMIT},
|
|
{VKD3D_SM5_OP_CUT_STREAM, VKD3D_STAT_CUT},
|
|
|
|
{VKD3D_SM4_OP_SAMPLE, VKD3D_STAT_SAMPLE},
|
|
{VKD3D_SM4_OP_SAMPLE_LOD, VKD3D_STAT_SAMPLE},
|
|
{VKD3D_SM5_OP_SAMPLE_LOD_S, VKD3D_STAT_SAMPLE},
|
|
{VKD3D_SM5_OP_SAMPLE_CL_S, VKD3D_STAT_SAMPLE},
|
|
{VKD3D_SM4_OP_SAMPLE_C, VKD3D_STAT_SAMPLE_C},
|
|
{VKD3D_SM4_OP_SAMPLE_C_LZ, VKD3D_STAT_SAMPLE_C},
|
|
{VKD3D_SM5_OP_SAMPLE_C_LZ_S, VKD3D_STAT_SAMPLE_C},
|
|
{VKD3D_SM5_OP_SAMPLE_C_CL_S, VKD3D_STAT_SAMPLE_C},
|
|
{VKD3D_SM4_OP_SAMPLE_GRAD, VKD3D_STAT_SAMPLE_GRAD},
|
|
{VKD3D_SM5_OP_SAMPLE_GRAD_CL_S, VKD3D_STAT_SAMPLE_GRAD},
|
|
{VKD3D_SM4_OP_SAMPLE_B, VKD3D_STAT_SAMPLE_BIAS},
|
|
{VKD3D_SM4_OP_GATHER4, VKD3D_STAT_GATHER},
|
|
{VKD3D_SM5_OP_GATHER4_PO, VKD3D_STAT_GATHER},
|
|
{VKD3D_SM4_OP_LOD, VKD3D_STAT_LOD},
|
|
|
|
{VKD3D_SM4_OP_LD, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM4_OP_LD2DMS, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_UAV_TYPED, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_RAW, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_STRUCTURED, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_S, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD2DMS_S, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_UAV_TYPED_S, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_RAW_S, VKD3D_STAT_LOAD},
|
|
{VKD3D_SM5_OP_LD_STRUCTURED_S, VKD3D_STAT_LOAD},
|
|
|
|
{VKD3D_SM5_OP_STORE_UAV_TYPED, VKD3D_STAT_STORE},
|
|
{VKD3D_SM5_OP_STORE_RAW, VKD3D_STAT_STORE},
|
|
{VKD3D_SM5_OP_STORE_STRUCTURED,VKD3D_STAT_STORE},
|
|
|
|
{VKD3D_SM4_OP_DCL_VERTICES_OUT, VKD3D_STAT_DCL_VERTICES_OUT},
|
|
{VKD3D_SM4_OP_DCL_INPUT_PRIMITIVE, VKD3D_STAT_DCL_INPUT_PRIMITIVE},
|
|
{VKD3D_SM4_OP_DCL_OUTPUT_TOPOLOGY, VKD3D_STAT_DCL_OUTPUT_TOPOLOGY},
|
|
{VKD3D_SM5_OP_DCL_GS_INSTANCES, VKD3D_STAT_DCL_GS_INSTANCES},
|
|
|
|
{VKD3D_SM4_OP_AND, VKD3D_STAT_BITWISE},
|
|
{VKD3D_SM4_OP_NOT, VKD3D_STAT_BITWISE},
|
|
{VKD3D_SM4_OP_OR, VKD3D_STAT_BITWISE},
|
|
{VKD3D_SM4_OP_XOR, VKD3D_STAT_BITWISE},
|
|
|
|
{VKD3D_SM5_OP_ATOMIC_AND, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_OR, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_XOR, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_CMP_STORE, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_IADD, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_IMAX, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_IMIN, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_UMAX, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_ATOMIC_UMIN, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_ALLOC, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_CONSUME, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IADD, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_AND, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_OR, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_XOR, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_EXCH, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_CMP_EXCH, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IMAX, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_IMIN, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_UMAX, VKD3D_STAT_ATOMIC},
|
|
{VKD3D_SM5_OP_IMM_ATOMIC_UMIN, VKD3D_STAT_ATOMIC},
|
|
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_DOMAIN, VKD3D_STAT_TESS_DOMAIN},
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_PARTITIONING, VKD3D_STAT_TESS_PARTITIONING},
|
|
{VKD3D_SM5_OP_DCL_TESSELLATOR_OUTPUT_PRIMITIVE, VKD3D_STAT_TESS_OUTPUT_PRIMITIVE},
|
|
{VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT, VKD3D_STAT_TESS_CONTROL_POINT_COUNT},
|
|
{VKD3D_SM5_OP_DCL_OUTPUT_CONTROL_POINT_COUNT, VKD3D_STAT_TESS_CONTROL_POINT_COUNT},
|
|
|
|
{VKD3D_SM5_OP_SYNC, VKD3D_STAT_BARRIER},
|
|
|
|
{VKD3D_SM4_OP_DCL_TEMPS, VKD3D_STAT_TEMPS},
|
|
};
|
|
|
|
memset(lookup, 0, sizeof(*lookup));
|
|
|
|
for (i = 0; i < ARRAY_SIZE(opcode_table); ++i)
|
|
{
|
|
const struct vkd3d_sm4_opcode_info *info = &opcode_table[i];
|
|
|
|
lookup->opcode_info_from_sm4[info->opcode] = info;
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(register_type_table); ++i)
|
|
{
|
|
const struct vkd3d_sm4_register_type_info *info = ®ister_type_table[i];
|
|
|
|
lookup->register_type_info_from_sm4[info->sm4_type] = info;
|
|
lookup->register_type_info_from_vkd3d[info->vkd3d_type] = info;
|
|
}
|
|
|
|
for (i = 0; i < ARRAY_SIZE(stat_field_table); ++i)
|
|
{
|
|
const struct vkd3d_sm4_stat_field_info *info = &stat_field_table[i];
|
|
|
|
lookup->stat_field_from_sm4[info->opcode] = info;
|
|
}
|
|
}
|
|
|
|
static void tpf_compiler_init(struct tpf_compiler *tpf, struct hlsl_ctx *ctx, struct sm4_stat *stat,
|
|
struct vkd3d_bytecode_buffer *buffer)
|
|
{
|
|
tpf->ctx = ctx;
|
|
tpf->buffer = buffer;
|
|
tpf->stat = stat;
|
|
init_sm4_lookup_tables(&tpf->lookup);
|
|
}
|
|
|
|
static const struct vkd3d_sm4_opcode_info *get_info_from_sm4_opcode(
|
|
const struct vkd3d_sm4_lookup_tables *lookup, enum vkd3d_sm4_opcode sm4_opcode)
|
|
{
|
|
if (sm4_opcode >= VKD3D_SM4_OP_COUNT)
|
|
return NULL;
|
|
return lookup->opcode_info_from_sm4[sm4_opcode];
|
|
}
|
|
|
|
static const struct vkd3d_sm4_register_type_info *get_info_from_sm4_register_type(
|
|
const struct vkd3d_sm4_lookup_tables *lookup, enum vkd3d_sm4_register_type sm4_type)
|
|
{
|
|
if (sm4_type >= VKD3D_SM4_REGISTER_TYPE_COUNT)
|
|
return NULL;
|
|
return lookup->register_type_info_from_sm4[sm4_type];
|
|
}
|
|
|
|
static const struct vkd3d_sm4_register_type_info *get_info_from_vkd3d_register_type(
|
|
const struct vkd3d_sm4_lookup_tables *lookup, enum vkd3d_shader_register_type vkd3d_type)
|
|
{
|
|
if (vkd3d_type >= VKD3DSPR_COUNT)
|
|
return NULL;
|
|
return lookup->register_type_info_from_vkd3d[vkd3d_type];
|
|
}
|
|
|
|
static enum vkd3d_sm4_swizzle_type vkd3d_sm4_get_default_swizzle_type(
|
|
const struct vkd3d_sm4_lookup_tables *lookup, enum vkd3d_shader_register_type vkd3d_type)
|
|
{
|
|
const struct vkd3d_sm4_register_type_info *register_type_info =
|
|
get_info_from_vkd3d_register_type(lookup, vkd3d_type);
|
|
|
|
VKD3D_ASSERT(register_type_info);
|
|
return register_type_info->default_src_swizzle_type;
|
|
}
|
|
|
|
static enum vkd3d_sm4_stat_field get_stat_field_from_sm4_opcode(
|
|
const struct vkd3d_sm4_lookup_tables *lookup, enum vkd3d_sm4_opcode sm4_opcode)
|
|
{
|
|
const struct vkd3d_sm4_stat_field_info *field_info;
|
|
|
|
if (sm4_opcode >= VKD3D_SM4_OP_COUNT || !(field_info = lookup->stat_field_from_sm4[sm4_opcode]))
|
|
return VKD3D_STAT_UNUSED;
|
|
return field_info->field;
|
|
}
|
|
|
|
static enum vkd3d_data_type map_data_type(char t)
|
|
{
|
|
switch (t)
|
|
{
|
|
case 'd':
|
|
return VKD3D_DATA_DOUBLE;
|
|
case 'f':
|
|
return VKD3D_DATA_FLOAT;
|
|
case 'i':
|
|
return VKD3D_DATA_INT;
|
|
case 'u':
|
|
return VKD3D_DATA_UINT;
|
|
case 'O':
|
|
return VKD3D_DATA_OPAQUE;
|
|
case '*':
|
|
return VKD3D_DATA_UNUSED;
|
|
default:
|
|
ERR("Invalid data type '%c'.\n", t);
|
|
return VKD3D_DATA_FLOAT;
|
|
}
|
|
}
|
|
|
|
static bool shader_sm4_read_reg_idx(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr,
|
|
const uint32_t *end, uint32_t addressing, struct vkd3d_shader_register_index *reg_idx)
|
|
{
|
|
if (addressing & VKD3D_SM4_ADDRESSING_RELATIVE)
|
|
{
|
|
struct vkd3d_shader_src_param *rel_addr = vsir_program_get_src_params(priv->p.program, 1);
|
|
|
|
if (!(reg_idx->rel_addr = rel_addr))
|
|
{
|
|
ERR("Failed to get src param for relative addressing.\n");
|
|
return false;
|
|
}
|
|
|
|
if (addressing & VKD3D_SM4_ADDRESSING_OFFSET)
|
|
reg_idx->offset = *(*ptr)++;
|
|
else
|
|
reg_idx->offset = 0;
|
|
shader_sm4_read_src_param(priv, ptr, end, VKD3D_DATA_INT, rel_addr);
|
|
}
|
|
else
|
|
{
|
|
reg_idx->rel_addr = NULL;
|
|
reg_idx->offset = *(*ptr)++;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool shader_sm4_read_param(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr, const uint32_t *end,
|
|
enum vkd3d_data_type data_type, struct vkd3d_shader_register *param, enum vkd3d_shader_src_modifier *modifier)
|
|
{
|
|
const struct vkd3d_sm4_register_type_info *register_type_info;
|
|
enum vkd3d_shader_register_type vsir_register_type;
|
|
enum vkd3d_sm4_register_precision precision;
|
|
uint32_t token, order, extended, addressing;
|
|
enum vkd3d_sm4_register_type register_type;
|
|
enum vkd3d_sm4_extended_operand_type type;
|
|
enum vkd3d_sm4_dimension sm4_dimension;
|
|
enum vkd3d_sm4_register_modifier m;
|
|
|
|
if (*ptr >= end)
|
|
{
|
|
WARN("Invalid ptr %p >= end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
token = *(*ptr)++;
|
|
|
|
register_type = (token & VKD3D_SM4_REGISTER_TYPE_MASK) >> VKD3D_SM4_REGISTER_TYPE_SHIFT;
|
|
register_type_info = get_info_from_sm4_register_type(&priv->lookup, register_type);
|
|
if (!register_type_info)
|
|
{
|
|
FIXME("Unhandled register type %#x.\n", register_type);
|
|
vsir_register_type = VKD3DSPR_TEMP;
|
|
}
|
|
else
|
|
{
|
|
vsir_register_type = register_type_info->vkd3d_type;
|
|
}
|
|
|
|
order = (token & VKD3D_SM4_REGISTER_ORDER_MASK) >> VKD3D_SM4_REGISTER_ORDER_SHIFT;
|
|
|
|
vsir_register_init(param, vsir_register_type, data_type, order);
|
|
param->precision = VKD3D_SHADER_REGISTER_PRECISION_DEFAULT;
|
|
param->non_uniform = false;
|
|
|
|
*modifier = VKD3DSPSM_NONE;
|
|
if (token & VKD3D_SM4_EXTENDED_OPERAND)
|
|
{
|
|
if (*ptr >= end)
|
|
{
|
|
WARN("Invalid ptr %p >= end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
extended = *(*ptr)++;
|
|
|
|
if (extended & VKD3D_SM4_EXTENDED_OPERAND)
|
|
{
|
|
FIXME("Skipping second-order extended operand.\n");
|
|
*ptr += *ptr < end;
|
|
}
|
|
|
|
type = extended & VKD3D_SM4_EXTENDED_OPERAND_TYPE_MASK;
|
|
if (type == VKD3D_SM4_EXTENDED_OPERAND_MODIFIER)
|
|
{
|
|
m = (extended & VKD3D_SM4_REGISTER_MODIFIER_MASK) >> VKD3D_SM4_REGISTER_MODIFIER_SHIFT;
|
|
switch (m)
|
|
{
|
|
case VKD3D_SM4_REGISTER_MODIFIER_NEGATE:
|
|
*modifier = VKD3DSPSM_NEG;
|
|
break;
|
|
|
|
case VKD3D_SM4_REGISTER_MODIFIER_ABS:
|
|
*modifier = VKD3DSPSM_ABS;
|
|
break;
|
|
|
|
case VKD3D_SM4_REGISTER_MODIFIER_ABS_NEGATE:
|
|
*modifier = VKD3DSPSM_ABSNEG;
|
|
break;
|
|
|
|
case VKD3D_SM4_REGISTER_MODIFIER_NONE:
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled register modifier %#x.\n", m);
|
|
break;
|
|
}
|
|
|
|
precision = (extended & VKD3D_SM4_REGISTER_PRECISION_MASK) >> VKD3D_SM4_REGISTER_PRECISION_SHIFT;
|
|
if (precision >= ARRAY_SIZE(register_precision_table)
|
|
|| register_precision_table[precision] == VKD3D_SHADER_REGISTER_PRECISION_INVALID)
|
|
{
|
|
FIXME("Unhandled register precision %#x.\n", precision);
|
|
param->precision = VKD3D_SHADER_REGISTER_PRECISION_INVALID;
|
|
}
|
|
else
|
|
{
|
|
param->precision = register_precision_table[precision];
|
|
}
|
|
|
|
if (extended & VKD3D_SM4_REGISTER_NON_UNIFORM_MASK)
|
|
param->non_uniform = true;
|
|
|
|
extended &= ~(VKD3D_SM4_EXTENDED_OPERAND_TYPE_MASK | VKD3D_SM4_REGISTER_MODIFIER_MASK
|
|
| VKD3D_SM4_REGISTER_PRECISION_MASK | VKD3D_SM4_REGISTER_NON_UNIFORM_MASK
|
|
| VKD3D_SM4_EXTENDED_OPERAND);
|
|
if (extended)
|
|
FIXME("Skipping unhandled extended operand bits 0x%08x.\n", extended);
|
|
}
|
|
else if (type)
|
|
{
|
|
FIXME("Skipping unhandled extended operand token 0x%08x (type %#x).\n", extended, type);
|
|
}
|
|
}
|
|
|
|
if (order >= 1)
|
|
{
|
|
addressing = (token & VKD3D_SM4_ADDRESSING_MASK0) >> VKD3D_SM4_ADDRESSING_SHIFT0;
|
|
if (!(shader_sm4_read_reg_idx(priv, ptr, end, addressing, ¶m->idx[0])))
|
|
{
|
|
ERR("Failed to read register index.\n");
|
|
return false;
|
|
}
|
|
}
|
|
|
|
if (order >= 2)
|
|
{
|
|
addressing = (token & VKD3D_SM4_ADDRESSING_MASK1) >> VKD3D_SM4_ADDRESSING_SHIFT1;
|
|
if (!(shader_sm4_read_reg_idx(priv, ptr, end, addressing, ¶m->idx[1])))
|
|
{
|
|
ERR("Failed to read register index.\n");
|
|
return false;
|
|
}
|
|
}
|
|
|
|
if (order >= 3)
|
|
{
|
|
addressing = (token & VKD3D_SM4_ADDRESSING_MASK2) >> VKD3D_SM4_ADDRESSING_SHIFT2;
|
|
if (!(shader_sm4_read_reg_idx(priv, ptr, end, addressing, ¶m->idx[2])))
|
|
{
|
|
ERR("Failed to read register index.\n");
|
|
return false;
|
|
}
|
|
}
|
|
|
|
if (order > 3)
|
|
{
|
|
WARN("Unhandled order %u.\n", order);
|
|
return false;
|
|
}
|
|
|
|
sm4_dimension = (token & VKD3D_SM4_DIMENSION_MASK) >> VKD3D_SM4_DIMENSION_SHIFT;
|
|
param->dimension = vsir_dimension_from_sm4_dimension(sm4_dimension);
|
|
|
|
if (register_type == VKD3D_SM4_RT_IMMCONST || register_type == VKD3D_SM4_RT_IMMCONST64)
|
|
{
|
|
unsigned int u32_count;
|
|
|
|
switch (param->dimension)
|
|
{
|
|
case VSIR_DIMENSION_SCALAR:
|
|
u32_count = 1 + (register_type == VKD3D_SM4_RT_IMMCONST64);
|
|
if (end - *ptr < u32_count)
|
|
{
|
|
WARN("Invalid ptr %p, end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
memcpy(param->u.immconst_u32, *ptr, u32_count * sizeof(uint32_t));
|
|
*ptr += u32_count;
|
|
break;
|
|
|
|
case VSIR_DIMENSION_VEC4:
|
|
if (end - *ptr < VKD3D_VEC4_SIZE)
|
|
{
|
|
WARN("Invalid ptr %p, end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
memcpy(param->u.immconst_u32, *ptr, VKD3D_VEC4_SIZE * sizeof(uint32_t));
|
|
*ptr += 4;
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled dimension %#x.\n", param->dimension);
|
|
break;
|
|
}
|
|
}
|
|
else if (register_type == VKD3D_SM4_RT_IMMCONSTBUFFER)
|
|
{
|
|
if (param->idx_count != 1)
|
|
{
|
|
WARN("Unexpected idx count %u.\n", param->idx_count);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_INDEX_COUNT,
|
|
"Invalid index count %u for immediate const buffer register; expected count 1.", param->idx_count);
|
|
}
|
|
}
|
|
else if (!shader_is_sm_5_1(priv) && vsir_register_is_descriptor(param))
|
|
{
|
|
/* SM5.1 places a symbol identifier in idx[0] and moves
|
|
* other values up one slot. Normalize to SM5.1. */
|
|
param->idx[2] = param->idx[1];
|
|
param->idx[1] = param->idx[0];
|
|
++param->idx_count;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool shader_sm4_is_scalar_register(const struct vkd3d_shader_register *reg)
|
|
{
|
|
switch (reg->type)
|
|
{
|
|
case VKD3DSPR_COVERAGE:
|
|
case VKD3DSPR_DEPTHOUT:
|
|
case VKD3DSPR_DEPTHOUTGE:
|
|
case VKD3DSPR_DEPTHOUTLE:
|
|
case VKD3DSPR_GSINSTID:
|
|
case VKD3DSPR_LOCALTHREADINDEX:
|
|
case VKD3DSPR_OUTPOINTID:
|
|
case VKD3DSPR_PRIMID:
|
|
case VKD3DSPR_SAMPLEMASK:
|
|
case VKD3DSPR_OUTSTENCILREF:
|
|
return true;
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static uint32_t swizzle_from_sm4(uint32_t s)
|
|
{
|
|
return vkd3d_shader_create_swizzle(s & 0x3, (s >> 2) & 0x3, (s >> 4) & 0x3, (s >> 6) & 0x3);
|
|
}
|
|
|
|
static uint32_t swizzle_to_sm4(uint32_t s)
|
|
{
|
|
uint32_t ret = 0;
|
|
ret |= ((vsir_swizzle_get_component(s, 0)) & 0x3);
|
|
ret |= ((vsir_swizzle_get_component(s, 1)) & 0x3) << 2;
|
|
ret |= ((vsir_swizzle_get_component(s, 2)) & 0x3) << 4;
|
|
ret |= ((vsir_swizzle_get_component(s, 3)) & 0x3) << 6;
|
|
return ret;
|
|
}
|
|
|
|
static bool register_is_input_output(const struct vkd3d_shader_register *reg)
|
|
{
|
|
switch (reg->type)
|
|
{
|
|
case VKD3DSPR_INPUT:
|
|
case VKD3DSPR_OUTPUT:
|
|
case VKD3DSPR_COLOROUT:
|
|
case VKD3DSPR_INCONTROLPOINT:
|
|
case VKD3DSPR_OUTCONTROLPOINT:
|
|
case VKD3DSPR_PATCHCONST:
|
|
return true;
|
|
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
static bool register_is_control_point_input(const struct vkd3d_shader_register *reg,
|
|
const struct vkd3d_shader_sm4_parser *priv)
|
|
{
|
|
return reg->type == VKD3DSPR_INCONTROLPOINT || reg->type == VKD3DSPR_OUTCONTROLPOINT
|
|
|| (reg->type == VKD3DSPR_INPUT && (priv->phase == VKD3DSIH_HS_CONTROL_POINT_PHASE
|
|
|| priv->p.program->shader_version.type == VKD3D_SHADER_TYPE_GEOMETRY));
|
|
}
|
|
|
|
static uint32_t mask_from_swizzle(uint32_t swizzle)
|
|
{
|
|
return (1u << vsir_swizzle_get_component(swizzle, 0))
|
|
| (1u << vsir_swizzle_get_component(swizzle, 1))
|
|
| (1u << vsir_swizzle_get_component(swizzle, 2))
|
|
| (1u << vsir_swizzle_get_component(swizzle, 3));
|
|
}
|
|
|
|
static bool shader_sm4_validate_input_output_register(struct vkd3d_shader_sm4_parser *priv,
|
|
const struct vkd3d_shader_register *reg, unsigned int mask)
|
|
{
|
|
unsigned int idx_count = 1 + register_is_control_point_input(reg, priv);
|
|
const unsigned int *masks;
|
|
unsigned int register_idx;
|
|
|
|
if (reg->idx_count != idx_count)
|
|
{
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_INDEX_COUNT,
|
|
"Invalid index count %u for register type %#x; expected count %u.",
|
|
reg->idx_count, reg->type, idx_count);
|
|
return false;
|
|
}
|
|
|
|
switch (reg->type)
|
|
{
|
|
case VKD3DSPR_INPUT:
|
|
case VKD3DSPR_INCONTROLPOINT:
|
|
masks = priv->input_register_masks;
|
|
break;
|
|
case VKD3DSPR_OUTPUT:
|
|
masks = sm4_parser_is_in_fork_or_join_phase(priv) ? priv->patch_constant_register_masks
|
|
: priv->output_register_masks;
|
|
break;
|
|
case VKD3DSPR_COLOROUT:
|
|
case VKD3DSPR_OUTCONTROLPOINT:
|
|
masks = priv->output_register_masks;
|
|
break;
|
|
case VKD3DSPR_PATCHCONST:
|
|
masks = priv->patch_constant_register_masks;
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
register_idx = reg->idx[reg->idx_count - 1].offset;
|
|
/* The signature element registers have already been checked against MAX_REG_OUTPUT. */
|
|
if (register_idx >= MAX_REG_OUTPUT || (masks[register_idx] & mask) != mask)
|
|
{
|
|
WARN("Failed to find signature element for register type %#x, index %u and mask %#x.\n",
|
|
reg->type, register_idx, mask);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_IO_REGISTER,
|
|
"Could not find signature element matching register type %#x, index %u and mask %#x.",
|
|
reg->type, register_idx, mask);
|
|
return false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool shader_sm4_read_src_param(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr,
|
|
const uint32_t *end, enum vkd3d_data_type data_type, struct vkd3d_shader_src_param *src_param)
|
|
{
|
|
unsigned int dimension, mask;
|
|
uint32_t token;
|
|
|
|
if (*ptr >= end)
|
|
{
|
|
WARN("Invalid ptr %p >= end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
token = **ptr;
|
|
|
|
if (!shader_sm4_read_param(priv, ptr, end, data_type, &src_param->reg, &src_param->modifiers))
|
|
{
|
|
ERR("Failed to read parameter.\n");
|
|
return false;
|
|
}
|
|
|
|
switch ((dimension = (token & VKD3D_SM4_DIMENSION_MASK) >> VKD3D_SM4_DIMENSION_SHIFT))
|
|
{
|
|
case VKD3D_SM4_DIMENSION_NONE:
|
|
case VKD3D_SM4_DIMENSION_SCALAR:
|
|
src_param->swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
break;
|
|
|
|
case VKD3D_SM4_DIMENSION_VEC4:
|
|
{
|
|
enum vkd3d_sm4_swizzle_type swizzle_type =
|
|
(token & VKD3D_SM4_SWIZZLE_TYPE_MASK) >> VKD3D_SM4_SWIZZLE_TYPE_SHIFT;
|
|
|
|
switch (swizzle_type)
|
|
{
|
|
case VKD3D_SM4_SWIZZLE_NONE:
|
|
src_param->swizzle = VKD3D_SHADER_NO_SWIZZLE;
|
|
|
|
mask = (token & VKD3D_SM4_WRITEMASK_MASK) >> VKD3D_SM4_WRITEMASK_SHIFT;
|
|
/* Mask seems only to be used for vec4 constants and is always zero. */
|
|
if (!register_is_constant(&src_param->reg))
|
|
{
|
|
FIXME("Source mask %#x is not for a constant.\n", mask);
|
|
vkd3d_shader_parser_warning(&priv->p, VKD3D_SHADER_WARNING_TPF_UNHANDLED_REGISTER_MASK,
|
|
"Unhandled mask %#x for a non-constant source register.", mask);
|
|
}
|
|
else if (mask)
|
|
{
|
|
FIXME("Unhandled mask %#x.\n", mask);
|
|
vkd3d_shader_parser_warning(&priv->p, VKD3D_SHADER_WARNING_TPF_UNHANDLED_REGISTER_MASK,
|
|
"Unhandled source register mask %#x.", mask);
|
|
}
|
|
|
|
break;
|
|
|
|
case VKD3D_SM4_SWIZZLE_SCALAR:
|
|
src_param->swizzle = (token & VKD3D_SM4_SWIZZLE_MASK) >> VKD3D_SM4_SWIZZLE_SHIFT;
|
|
src_param->swizzle = (src_param->swizzle & 0x3) * 0x01010101;
|
|
break;
|
|
|
|
case VKD3D_SM4_SWIZZLE_VEC4:
|
|
src_param->swizzle = swizzle_from_sm4((token & VKD3D_SM4_SWIZZLE_MASK) >> VKD3D_SM4_SWIZZLE_SHIFT);
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled swizzle type %#x.\n", swizzle_type);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_SWIZZLE,
|
|
"Source register swizzle type %#x is invalid.", swizzle_type);
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
default:
|
|
FIXME("Unhandled dimension %#x.\n", dimension);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_DIMENSION,
|
|
"Source register dimension %#x is invalid.", dimension);
|
|
break;
|
|
}
|
|
|
|
if (data_type_is_64_bit(data_type))
|
|
src_param->swizzle = vsir_swizzle_64_from_32(src_param->swizzle);
|
|
|
|
if (register_is_input_output(&src_param->reg) && !shader_sm4_validate_input_output_register(priv,
|
|
&src_param->reg, mask_from_swizzle(src_param->swizzle)))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool shader_sm4_read_dst_param(struct vkd3d_shader_sm4_parser *priv, const uint32_t **ptr,
|
|
const uint32_t *end, enum vkd3d_data_type data_type, struct vkd3d_shader_dst_param *dst_param)
|
|
{
|
|
enum vkd3d_sm4_swizzle_type swizzle_type;
|
|
enum vkd3d_shader_src_modifier modifier;
|
|
unsigned int dimension, swizzle;
|
|
uint32_t token;
|
|
|
|
if (*ptr >= end)
|
|
{
|
|
WARN("Invalid ptr %p >= end %p.\n", *ptr, end);
|
|
return false;
|
|
}
|
|
token = **ptr;
|
|
|
|
if (!shader_sm4_read_param(priv, ptr, end, data_type, &dst_param->reg, &modifier))
|
|
{
|
|
ERR("Failed to read parameter.\n");
|
|
return false;
|
|
}
|
|
|
|
if (modifier != VKD3DSPSM_NONE)
|
|
{
|
|
ERR("Invalid source modifier %#x on destination register.\n", modifier);
|
|
return false;
|
|
}
|
|
|
|
switch ((dimension = (token & VKD3D_SM4_DIMENSION_MASK) >> VKD3D_SM4_DIMENSION_SHIFT))
|
|
{
|
|
case VKD3D_SM4_DIMENSION_NONE:
|
|
dst_param->write_mask = 0;
|
|
break;
|
|
|
|
case VKD3D_SM4_DIMENSION_SCALAR:
|
|
dst_param->write_mask = VKD3DSP_WRITEMASK_0;
|
|
break;
|
|
|
|
case VKD3D_SM4_DIMENSION_VEC4:
|
|
swizzle_type = (token & VKD3D_SM4_SWIZZLE_TYPE_MASK) >> VKD3D_SM4_SWIZZLE_TYPE_SHIFT;
|
|
switch (swizzle_type)
|
|
{
|
|
case VKD3D_SM4_SWIZZLE_NONE:
|
|
dst_param->write_mask = (token & VKD3D_SM4_WRITEMASK_MASK) >> VKD3D_SM4_WRITEMASK_SHIFT;
|
|
break;
|
|
|
|
case VKD3D_SM4_SWIZZLE_VEC4:
|
|
swizzle = swizzle_from_sm4((token & VKD3D_SM4_SWIZZLE_MASK) >> VKD3D_SM4_SWIZZLE_SHIFT);
|
|
if (swizzle != VKD3D_SHADER_NO_SWIZZLE)
|
|
{
|
|
FIXME("Unhandled swizzle %#x.\n", swizzle);
|
|
vkd3d_shader_parser_warning(&priv->p, VKD3D_SHADER_WARNING_TPF_UNHANDLED_REGISTER_SWIZZLE,
|
|
"Unhandled destination register swizzle %#x.", swizzle);
|
|
}
|
|
dst_param->write_mask = VKD3DSP_WRITEMASK_ALL;
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled swizzle type %#x.\n", swizzle_type);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_SWIZZLE,
|
|
"Destination register swizzle type %#x is invalid.", swizzle_type);
|
|
break;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unhandled dimension %#x.\n", dimension);
|
|
vkd3d_shader_parser_error(&priv->p, VKD3D_SHADER_ERROR_TPF_INVALID_REGISTER_DIMENSION,
|
|
"Destination register dimension %#x is invalid.", dimension);
|
|
break;
|
|
}
|
|
|
|
if (data_type == VKD3D_DATA_DOUBLE)
|
|
dst_param->write_mask = vsir_write_mask_64_from_32(dst_param->write_mask);
|
|
/* Some scalar registers are declared with no write mask in shader bytecode. */
|
|
if (!dst_param->write_mask && shader_sm4_is_scalar_register(&dst_param->reg))
|
|
dst_param->write_mask = VKD3DSP_WRITEMASK_0;
|
|
dst_param->modifiers = 0;
|
|
dst_param->shift = 0;
|
|
|
|
if (register_is_input_output(&dst_param->reg) && !shader_sm4_validate_input_output_register(priv,
|
|
&dst_param->reg, dst_param->write_mask))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
static void shader_sm4_read_instruction_modifier(uint32_t modifier, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
enum vkd3d_sm4_instruction_modifier modifier_type = modifier & VKD3D_SM4_MODIFIER_MASK;
|
|
|
|
switch (modifier_type)
|
|
{
|
|
case VKD3D_SM4_MODIFIER_AOFFIMMI:
|
|
{
|
|
static const uint32_t recognized_bits = VKD3D_SM4_INSTRUCTION_MODIFIER
|
|
| VKD3D_SM4_MODIFIER_MASK
|
|
| VKD3D_SM4_AOFFIMMI_U_MASK
|
|
| VKD3D_SM4_AOFFIMMI_V_MASK
|
|
| VKD3D_SM4_AOFFIMMI_W_MASK;
|
|
|
|
/* Bit fields are used for sign extension. */
|
|
struct
|
|
{
|
|
int u : 4;
|
|
int v : 4;
|
|
int w : 4;
|
|
} aoffimmi;
|
|
|
|
if (modifier & ~recognized_bits)
|
|
FIXME("Unhandled instruction modifier %#x.\n", modifier);
|
|
|
|
aoffimmi.u = (modifier & VKD3D_SM4_AOFFIMMI_U_MASK) >> VKD3D_SM4_AOFFIMMI_U_SHIFT;
|
|
aoffimmi.v = (modifier & VKD3D_SM4_AOFFIMMI_V_MASK) >> VKD3D_SM4_AOFFIMMI_V_SHIFT;
|
|
aoffimmi.w = (modifier & VKD3D_SM4_AOFFIMMI_W_MASK) >> VKD3D_SM4_AOFFIMMI_W_SHIFT;
|
|
ins->texel_offset.u = aoffimmi.u;
|
|
ins->texel_offset.v = aoffimmi.v;
|
|
ins->texel_offset.w = aoffimmi.w;
|
|
break;
|
|
}
|
|
|
|
case VKD3D_SM5_MODIFIER_DATA_TYPE:
|
|
{
|
|
uint32_t components = (modifier & VKD3D_SM5_MODIFIER_DATA_TYPE_MASK) >> VKD3D_SM5_MODIFIER_DATA_TYPE_SHIFT;
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < VKD3D_VEC4_SIZE; i++)
|
|
{
|
|
enum vkd3d_sm4_data_type data_type = VKD3D_SM4_TYPE_COMPONENT(components, i);
|
|
|
|
if (!data_type || (data_type >= ARRAY_SIZE(data_type_table)))
|
|
{
|
|
FIXME("Unhandled data type %#x.\n", data_type);
|
|
ins->resource_data_type[i] = VKD3D_DATA_FLOAT;
|
|
}
|
|
else
|
|
{
|
|
ins->resource_data_type[i] = data_type_table[data_type];
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
|
|
case VKD3D_SM5_MODIFIER_RESOURCE_TYPE:
|
|
{
|
|
enum vkd3d_sm4_resource_type resource_type
|
|
= (modifier & VKD3D_SM5_MODIFIER_RESOURCE_TYPE_MASK) >> VKD3D_SM5_MODIFIER_RESOURCE_TYPE_SHIFT;
|
|
|
|
if (resource_type == VKD3D_SM4_RESOURCE_RAW_BUFFER)
|
|
ins->raw = true;
|
|
else if (resource_type == VKD3D_SM4_RESOURCE_STRUCTURED_BUFFER)
|
|
ins->structured = true;
|
|
|
|
if (resource_type < ARRAY_SIZE(resource_type_table))
|
|
ins->resource_type = resource_type_table[resource_type];
|
|
else
|
|
{
|
|
FIXME("Unhandled resource type %#x.\n", resource_type);
|
|
ins->resource_type = VKD3D_SHADER_RESOURCE_NONE;
|
|
}
|
|
|
|
ins->resource_stride
|
|
= (modifier & VKD3D_SM5_MODIFIER_RESOURCE_STRIDE_MASK) >> VKD3D_SM5_MODIFIER_RESOURCE_STRIDE_SHIFT;
|
|
break;
|
|
}
|
|
|
|
default:
|
|
FIXME("Unhandled instruction modifier %#x.\n", modifier);
|
|
}
|
|
}
|
|
|
|
static void shader_sm4_read_instruction(struct vkd3d_shader_sm4_parser *sm4, struct vkd3d_shader_instruction *ins)
|
|
{
|
|
const struct vkd3d_sm4_opcode_info *opcode_info;
|
|
struct vsir_program *program = sm4->p.program;
|
|
uint32_t opcode_token, opcode, previous_token;
|
|
struct vkd3d_shader_dst_param *dst_params;
|
|
struct vkd3d_shader_src_param *src_params;
|
|
const uint32_t **ptr = &sm4->ptr;
|
|
unsigned int i, len;
|
|
const uint32_t *p;
|
|
uint32_t precise;
|
|
size_t remaining;
|
|
|
|
if (*ptr >= sm4->end)
|
|
{
|
|
WARN("End of byte-code, failed to read opcode.\n");
|
|
goto fail;
|
|
}
|
|
remaining = sm4->end - *ptr;
|
|
|
|
++sm4->p.location.line;
|
|
|
|
opcode_token = *(*ptr)++;
|
|
opcode = opcode_token & VKD3D_SM4_OPCODE_MASK;
|
|
|
|
len = ((opcode_token & VKD3D_SM4_INSTRUCTION_LENGTH_MASK) >> VKD3D_SM4_INSTRUCTION_LENGTH_SHIFT);
|
|
if (!len)
|
|
{
|
|
if (remaining < 2)
|
|
{
|
|
WARN("End of byte-code, failed to read length token.\n");
|
|
goto fail;
|
|
}
|
|
len = **ptr;
|
|
}
|
|
if (!len || remaining < len)
|
|
{
|
|
WARN("Read invalid length %u (remaining %zu).\n", len, remaining);
|
|
goto fail;
|
|
}
|
|
--len;
|
|
|
|
if (!(opcode_info = get_info_from_sm4_opcode(&sm4->lookup, opcode)))
|
|
{
|
|
FIXME("Unrecognized opcode %#x, opcode_token 0x%08x.\n", opcode, opcode_token);
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
*ptr += len;
|
|
return;
|
|
}
|
|
|
|
vsir_instruction_init(ins, &sm4->p.location, opcode_info->handler_idx);
|
|
if (ins->opcode == VKD3DSIH_HS_CONTROL_POINT_PHASE || ins->opcode == VKD3DSIH_HS_FORK_PHASE
|
|
|| ins->opcode == VKD3DSIH_HS_JOIN_PHASE)
|
|
sm4->phase = ins->opcode;
|
|
sm4->has_control_point_phase |= ins->opcode == VKD3DSIH_HS_CONTROL_POINT_PHASE;
|
|
ins->flags = 0;
|
|
ins->coissue = false;
|
|
ins->raw = false;
|
|
ins->structured = false;
|
|
ins->predicate = NULL;
|
|
ins->dst_count = strnlen(opcode_info->dst_info, SM4_MAX_DST_COUNT);
|
|
ins->src_count = strnlen(opcode_info->src_info, SM4_MAX_SRC_COUNT);
|
|
ins->src = src_params = vsir_program_get_src_params(program, ins->src_count);
|
|
if (!src_params && ins->src_count)
|
|
{
|
|
ERR("Failed to allocate src parameters.\n");
|
|
vkd3d_shader_parser_error(&sm4->p, VKD3D_SHADER_ERROR_TPF_OUT_OF_MEMORY, "Out of memory.");
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
ins->resource_type = VKD3D_SHADER_RESOURCE_NONE;
|
|
ins->resource_stride = 0;
|
|
ins->resource_data_type[0] = VKD3D_DATA_FLOAT;
|
|
ins->resource_data_type[1] = VKD3D_DATA_FLOAT;
|
|
ins->resource_data_type[2] = VKD3D_DATA_FLOAT;
|
|
ins->resource_data_type[3] = VKD3D_DATA_FLOAT;
|
|
memset(&ins->texel_offset, 0, sizeof(ins->texel_offset));
|
|
|
|
p = *ptr;
|
|
*ptr += len;
|
|
|
|
if (opcode_info->read_opcode_func)
|
|
{
|
|
ins->dst = NULL;
|
|
ins->dst_count = 0;
|
|
opcode_info->read_opcode_func(ins, opcode, opcode_token, p, len, sm4);
|
|
}
|
|
else
|
|
{
|
|
enum vkd3d_shader_dst_modifier instruction_dst_modifier = VKD3DSPDM_NONE;
|
|
|
|
previous_token = opcode_token;
|
|
while (previous_token & VKD3D_SM4_INSTRUCTION_MODIFIER && p != *ptr)
|
|
shader_sm4_read_instruction_modifier(previous_token = *p++, ins);
|
|
|
|
ins->flags = (opcode_token & VKD3D_SM4_INSTRUCTION_FLAGS_MASK) >> VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT;
|
|
if (ins->flags & VKD3D_SM4_INSTRUCTION_FLAG_SATURATE)
|
|
{
|
|
ins->flags &= ~VKD3D_SM4_INSTRUCTION_FLAG_SATURATE;
|
|
instruction_dst_modifier = VKD3DSPDM_SATURATE;
|
|
}
|
|
precise = (opcode_token & VKD3D_SM5_PRECISE_MASK) >> VKD3D_SM5_PRECISE_SHIFT;
|
|
ins->flags |= precise << VKD3DSI_PRECISE_SHIFT;
|
|
|
|
ins->dst = dst_params = vsir_program_get_dst_params(program, ins->dst_count);
|
|
if (!dst_params && ins->dst_count)
|
|
{
|
|
ERR("Failed to allocate dst parameters.\n");
|
|
vkd3d_shader_parser_error(&sm4->p, VKD3D_SHADER_ERROR_TPF_OUT_OF_MEMORY, "Out of memory.");
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
for (i = 0; i < ins->dst_count; ++i)
|
|
{
|
|
if (!(shader_sm4_read_dst_param(sm4, &p, *ptr, map_data_type(opcode_info->dst_info[i]),
|
|
&dst_params[i])))
|
|
{
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
dst_params[i].modifiers |= instruction_dst_modifier;
|
|
}
|
|
|
|
for (i = 0; i < ins->src_count; ++i)
|
|
{
|
|
if (!(shader_sm4_read_src_param(sm4, &p, *ptr, map_data_type(opcode_info->src_info[i]),
|
|
&src_params[i])))
|
|
{
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
return;
|
|
|
|
fail:
|
|
*ptr = sm4->end;
|
|
ins->opcode = VKD3DSIH_INVALID;
|
|
return;
|
|
}
|
|
|
|
static bool shader_sm4_init(struct vkd3d_shader_sm4_parser *sm4, struct vsir_program *program,
|
|
const uint32_t *byte_code, size_t byte_code_size, const struct vkd3d_shader_compile_info *compile_info,
|
|
struct vkd3d_shader_message_context *message_context)
|
|
{
|
|
struct vkd3d_shader_version version;
|
|
uint32_t version_token, token_count;
|
|
|
|
if (byte_code_size / sizeof(*byte_code) < 2)
|
|
{
|
|
WARN("Invalid byte code size %lu.\n", (long)byte_code_size);
|
|
return false;
|
|
}
|
|
|
|
version_token = byte_code[0];
|
|
TRACE("Version: 0x%08x.\n", version_token);
|
|
token_count = byte_code[1];
|
|
TRACE("Token count: %u.\n", token_count);
|
|
|
|
if (token_count < 2 || byte_code_size / sizeof(*byte_code) < token_count)
|
|
{
|
|
WARN("Invalid token count %u.\n", token_count);
|
|
return false;
|
|
}
|
|
|
|
sm4->start = &byte_code[2];
|
|
sm4->end = &byte_code[token_count];
|
|
|
|
switch (version_token >> 16)
|
|
{
|
|
case VKD3D_SM4_PS:
|
|
version.type = VKD3D_SHADER_TYPE_PIXEL;
|
|
break;
|
|
|
|
case VKD3D_SM4_VS:
|
|
version.type = VKD3D_SHADER_TYPE_VERTEX;
|
|
break;
|
|
|
|
case VKD3D_SM4_GS:
|
|
version.type = VKD3D_SHADER_TYPE_GEOMETRY;
|
|
break;
|
|
|
|
case VKD3D_SM5_HS:
|
|
version.type = VKD3D_SHADER_TYPE_HULL;
|
|
break;
|
|
|
|
case VKD3D_SM5_DS:
|
|
version.type = VKD3D_SHADER_TYPE_DOMAIN;
|
|
break;
|
|
|
|
case VKD3D_SM5_CS:
|
|
version.type = VKD3D_SHADER_TYPE_COMPUTE;
|
|
break;
|
|
|
|
default:
|
|
FIXME("Unrecognised shader type %#x.\n", version_token >> 16);
|
|
}
|
|
version.major = VKD3D_SM4_VERSION_MAJOR(version_token);
|
|
version.minor = VKD3D_SM4_VERSION_MINOR(version_token);
|
|
|
|
/* Estimate instruction count to avoid reallocation in most shaders. */
|
|
if (!vsir_program_init(program, compile_info, &version, token_count / 7u + 20, VSIR_CF_STRUCTURED))
|
|
return false;
|
|
vkd3d_shader_parser_init(&sm4->p, program, message_context, compile_info->source_name);
|
|
sm4->ptr = sm4->start;
|
|
|
|
init_sm4_lookup_tables(&sm4->lookup);
|
|
|
|
return true;
|
|
}
|
|
|
|
static void uninvert_used_masks(struct shader_signature *signature)
|
|
{
|
|
for (unsigned int i = 0; i < signature->element_count; ++i)
|
|
{
|
|
struct signature_element *e = &signature->elements[i];
|
|
|
|
e->used_mask = e->mask & ~e->used_mask;
|
|
}
|
|
}
|
|
|
|
static bool shader_sm4_parser_validate_signature(struct vkd3d_shader_sm4_parser *sm4,
|
|
const struct shader_signature *signature, unsigned int *masks, const char *name)
|
|
{
|
|
unsigned int i, register_idx, register_count, mask;
|
|
|
|
for (i = 0; i < signature->element_count; ++i)
|
|
{
|
|
register_idx = signature->elements[i].register_index;
|
|
register_count = signature->elements[i].register_count;
|
|
if (register_idx != ~0u && (register_idx >= MAX_REG_OUTPUT || MAX_REG_OUTPUT - register_idx < register_count))
|
|
{
|
|
WARN("%s signature element %u unhandled register index %u, count %u.\n",
|
|
name, i, register_idx, register_count);
|
|
vkd3d_shader_parser_error(&sm4->p, VKD3D_SHADER_ERROR_TPF_TOO_MANY_REGISTERS,
|
|
"%s signature element %u register index %u, count %u exceeds maximum index of %u.", name,
|
|
i, register_idx, register_count, MAX_REG_OUTPUT - 1);
|
|
return false;
|
|
}
|
|
|
|
if (!vkd3d_bitmask_is_contiguous(mask = signature->elements[i].mask))
|
|
{
|
|
WARN("%s signature element %u mask %#x is not contiguous.\n", name, i, mask);
|
|
vkd3d_shader_parser_warning(&sm4->p, VKD3D_SHADER_WARNING_TPF_MASK_NOT_CONTIGUOUS,
|
|
"%s signature element %u mask %#x is not contiguous.", name, i, mask);
|
|
}
|
|
|
|
if (register_idx != ~0u)
|
|
masks[register_idx] |= mask;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static int index_range_compare(const void *a, const void *b)
|
|
{
|
|
return memcmp(a, b, sizeof(struct sm4_index_range));
|
|
}
|
|
|
|
static void shader_sm4_validate_default_phase_index_ranges(struct vkd3d_shader_sm4_parser *sm4)
|
|
{
|
|
if (!sm4->input_index_ranges.count || !sm4->output_index_ranges.count)
|
|
return;
|
|
|
|
if (sm4->input_index_ranges.count == sm4->output_index_ranges.count)
|
|
{
|
|
qsort(sm4->input_index_ranges.ranges, sm4->input_index_ranges.count, sizeof(sm4->input_index_ranges.ranges[0]),
|
|
index_range_compare);
|
|
qsort(sm4->output_index_ranges.ranges, sm4->output_index_ranges.count, sizeof(sm4->output_index_ranges.ranges[0]),
|
|
index_range_compare);
|
|
if (!memcmp(sm4->input_index_ranges.ranges, sm4->output_index_ranges.ranges,
|
|
sm4->input_index_ranges.count * sizeof(sm4->input_index_ranges.ranges[0])))
|
|
return;
|
|
}
|
|
|
|
/* This is very unlikely to occur and would complicate the default control point phase implementation. */
|
|
WARN("Default phase index ranges are not identical.\n");
|
|
vkd3d_shader_parser_error(&sm4->p, VKD3D_SHADER_ERROR_TPF_INVALID_INDEX_RANGE_DCL,
|
|
"Default control point phase input and output index range declarations are not identical.");
|
|
return;
|
|
}
|
|
|
|
int tpf_parse(const struct vkd3d_shader_compile_info *compile_info, uint64_t config_flags,
|
|
struct vkd3d_shader_message_context *message_context, struct vsir_program *program)
|
|
{
|
|
struct vkd3d_shader_instruction_array *instructions;
|
|
struct vkd3d_shader_sm4_parser sm4 = {0};
|
|
struct dxbc_shader_desc dxbc_desc = {0};
|
|
struct vkd3d_shader_instruction *ins;
|
|
int ret;
|
|
|
|
dxbc_desc.is_dxil = false;
|
|
if ((ret = shader_extract_from_dxbc(&compile_info->source,
|
|
message_context, compile_info->source_name, &dxbc_desc)) < 0)
|
|
{
|
|
WARN("Failed to extract shader, vkd3d result %d.\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
if (!shader_sm4_init(&sm4, program, dxbc_desc.byte_code, dxbc_desc.byte_code_size,
|
|
compile_info, message_context))
|
|
{
|
|
WARN("Failed to initialise shader parser.\n");
|
|
free_dxbc_shader_desc(&dxbc_desc);
|
|
return VKD3D_ERROR_INVALID_ARGUMENT;
|
|
}
|
|
|
|
program->input_signature = dxbc_desc.input_signature;
|
|
program->output_signature = dxbc_desc.output_signature;
|
|
program->patch_constant_signature = dxbc_desc.patch_constant_signature;
|
|
memset(&dxbc_desc, 0, sizeof(dxbc_desc));
|
|
|
|
/* DXBC stores used masks inverted for output signatures, for some reason.
|
|
* We return them un-inverted. */
|
|
uninvert_used_masks(&program->output_signature);
|
|
if (program->shader_version.type == VKD3D_SHADER_TYPE_HULL)
|
|
uninvert_used_masks(&program->patch_constant_signature);
|
|
|
|
switch (program->shader_version.type)
|
|
{
|
|
case VKD3D_SHADER_TYPE_HULL:
|
|
case VKD3D_SHADER_TYPE_DOMAIN:
|
|
break;
|
|
|
|
default:
|
|
if (program->patch_constant_signature.element_count != 0)
|
|
{
|
|
WARN("The patch constant signature only makes sense for Hull and Domain Shaders, ignoring it.\n");
|
|
shader_signature_cleanup(&program->patch_constant_signature);
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (!shader_sm4_parser_validate_signature(&sm4, &program->input_signature,
|
|
sm4.input_register_masks, "Input")
|
|
|| !shader_sm4_parser_validate_signature(&sm4, &program->output_signature,
|
|
sm4.output_register_masks, "Output")
|
|
|| !shader_sm4_parser_validate_signature(&sm4, &program->patch_constant_signature,
|
|
sm4.patch_constant_register_masks, "Patch constant"))
|
|
{
|
|
vsir_program_cleanup(program);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
instructions = &program->instructions;
|
|
while (sm4.ptr != sm4.end)
|
|
{
|
|
if (!shader_instruction_array_reserve(instructions, instructions->count + 1))
|
|
{
|
|
ERR("Failed to allocate instructions.\n");
|
|
vkd3d_shader_parser_error(&sm4.p, VKD3D_SHADER_ERROR_TPF_OUT_OF_MEMORY, "Out of memory.");
|
|
vsir_program_cleanup(program);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
ins = &instructions->elements[instructions->count];
|
|
shader_sm4_read_instruction(&sm4, ins);
|
|
|
|
if (ins->opcode == VKD3DSIH_INVALID)
|
|
{
|
|
WARN("Encountered unrecognized or invalid instruction.\n");
|
|
vsir_program_cleanup(program);
|
|
return VKD3D_ERROR_OUT_OF_MEMORY;
|
|
}
|
|
++instructions->count;
|
|
}
|
|
if (program->shader_version.type == VKD3D_SHADER_TYPE_HULL
|
|
&& !sm4.has_control_point_phase && !sm4.p.failed)
|
|
shader_sm4_validate_default_phase_index_ranges(&sm4);
|
|
|
|
if (sm4.p.failed)
|
|
{
|
|
WARN("Failed to parse shader.\n");
|
|
vsir_program_cleanup(program);
|
|
return VKD3D_ERROR_INVALID_SHADER;
|
|
}
|
|
|
|
if ((ret = vkd3d_shader_parser_validate(&sm4.p, config_flags)) < 0)
|
|
{
|
|
WARN("Failed to validate shader after parsing, ret %d.\n", ret);
|
|
|
|
if (TRACE_ON())
|
|
vsir_program_trace(program);
|
|
|
|
vsir_program_cleanup(program);
|
|
return ret;
|
|
}
|
|
|
|
return VKD3D_OK;
|
|
}
|
|
|
|
static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_block *block);
|
|
|
|
static bool type_is_integer(const struct hlsl_type *type)
|
|
{
|
|
switch (type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
return true;
|
|
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
bool hlsl_sm4_register_from_semantic(struct hlsl_ctx *ctx, const struct hlsl_semantic *semantic,
|
|
bool output, enum vkd3d_shader_register_type *type, bool *has_idx)
|
|
{
|
|
unsigned int i;
|
|
|
|
static const struct
|
|
{
|
|
const char *semantic;
|
|
bool output;
|
|
enum vkd3d_shader_type shader_type;
|
|
enum vkd3d_shader_register_type type;
|
|
bool has_idx;
|
|
}
|
|
register_table[] =
|
|
{
|
|
{"sv_dispatchthreadid", false, VKD3D_SHADER_TYPE_COMPUTE, VKD3DSPR_THREADID, false},
|
|
{"sv_groupid", false, VKD3D_SHADER_TYPE_COMPUTE, VKD3DSPR_THREADGROUPID, false},
|
|
{"sv_groupthreadid", false, VKD3D_SHADER_TYPE_COMPUTE, VKD3DSPR_LOCALTHREADID, false},
|
|
|
|
{"sv_primitiveid", false, VKD3D_SHADER_TYPE_GEOMETRY, VKD3DSPR_PRIMID, false},
|
|
|
|
/* Put sv_target in this table, instead of letting it fall through to
|
|
* default varying allocation, so that the register index matches the
|
|
* usage index. */
|
|
{"color", true, VKD3D_SHADER_TYPE_PIXEL, VKD3DSPR_OUTPUT, true},
|
|
{"depth", true, VKD3D_SHADER_TYPE_PIXEL, VKD3DSPR_DEPTHOUT, false},
|
|
{"sv_depth", true, VKD3D_SHADER_TYPE_PIXEL, VKD3DSPR_DEPTHOUT, false},
|
|
{"sv_target", true, VKD3D_SHADER_TYPE_PIXEL, VKD3DSPR_OUTPUT, true},
|
|
{"sv_coverage", true, VKD3D_SHADER_TYPE_PIXEL, VKD3DSPR_SAMPLEMASK, false},
|
|
};
|
|
|
|
for (i = 0; i < ARRAY_SIZE(register_table); ++i)
|
|
{
|
|
if (!ascii_strcasecmp(semantic->name, register_table[i].semantic)
|
|
&& output == register_table[i].output
|
|
&& ctx->profile->type == register_table[i].shader_type)
|
|
{
|
|
if (type)
|
|
*type = register_table[i].type;
|
|
*has_idx = register_table[i].has_idx;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
bool sysval_semantic_from_hlsl(enum vkd3d_shader_sysval_semantic *semantic,
|
|
struct hlsl_ctx *ctx, const struct hlsl_semantic *hlsl_semantic, bool output)
|
|
{
|
|
unsigned int i;
|
|
|
|
static const struct
|
|
{
|
|
const char *name;
|
|
bool output;
|
|
enum vkd3d_shader_type shader_type;
|
|
enum vkd3d_shader_sysval_semantic semantic;
|
|
}
|
|
semantics[] =
|
|
{
|
|
{"sv_dispatchthreadid", false, VKD3D_SHADER_TYPE_COMPUTE, ~0u},
|
|
{"sv_groupid", false, VKD3D_SHADER_TYPE_COMPUTE, ~0u},
|
|
{"sv_groupthreadid", false, VKD3D_SHADER_TYPE_COMPUTE, ~0u},
|
|
|
|
{"position", false, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_position", false, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_primitiveid", false, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_PRIMITIVE_ID},
|
|
|
|
{"position", true, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_position", true, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_primitiveid", true, VKD3D_SHADER_TYPE_GEOMETRY, VKD3D_SHADER_SV_PRIMITIVE_ID},
|
|
|
|
{"position", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_position", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_primitiveid", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_PRIMITIVE_ID},
|
|
{"sv_isfrontface", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_IS_FRONT_FACE},
|
|
{"sv_rendertargetarrayindex", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_RENDER_TARGET_ARRAY_INDEX},
|
|
{"sv_viewportarrayindex", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_VIEWPORT_ARRAY_INDEX},
|
|
{"sv_sampleindex", false, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_SAMPLE_INDEX},
|
|
|
|
{"color", true, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_TARGET},
|
|
{"depth", true, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_DEPTH},
|
|
{"sv_target", true, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_TARGET},
|
|
{"sv_depth", true, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_DEPTH},
|
|
{"sv_coverage", true, VKD3D_SHADER_TYPE_PIXEL, VKD3D_SHADER_SV_COVERAGE},
|
|
|
|
{"sv_position", false, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_NONE},
|
|
{"sv_vertexid", false, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_VERTEX_ID},
|
|
{"sv_instanceid", false, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_INSTANCE_ID},
|
|
|
|
{"position", true, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_position", true, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_POSITION},
|
|
{"sv_rendertargetarrayindex", true, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_RENDER_TARGET_ARRAY_INDEX},
|
|
{"sv_viewportarrayindex", true, VKD3D_SHADER_TYPE_VERTEX, VKD3D_SHADER_SV_VIEWPORT_ARRAY_INDEX},
|
|
};
|
|
bool needs_compat_mapping = ascii_strncasecmp(hlsl_semantic->name, "sv_", 3);
|
|
|
|
for (i = 0; i < ARRAY_SIZE(semantics); ++i)
|
|
{
|
|
if (!ascii_strcasecmp(hlsl_semantic->name, semantics[i].name)
|
|
&& output == semantics[i].output
|
|
&& (ctx->semantic_compat_mapping == needs_compat_mapping || !needs_compat_mapping)
|
|
&& ctx->profile->type == semantics[i].shader_type)
|
|
{
|
|
*semantic = semantics[i].semantic;
|
|
return true;
|
|
}
|
|
}
|
|
|
|
if (!needs_compat_mapping)
|
|
return false;
|
|
|
|
*semantic = VKD3D_SHADER_SV_NONE;
|
|
return true;
|
|
}
|
|
|
|
static void add_section(struct hlsl_ctx *ctx, struct dxbc_writer *dxbc,
|
|
uint32_t tag, struct vkd3d_bytecode_buffer *buffer)
|
|
{
|
|
/* Native D3DDisassemble() expects at least the sizes of the ISGN and OSGN
|
|
* sections to be aligned. Without this, the sections themselves will be
|
|
* aligned, but their reported sizes won't. */
|
|
size_t size = bytecode_align(buffer);
|
|
|
|
dxbc_writer_add_section(dxbc, tag, buffer->data, size);
|
|
|
|
if (buffer->status < 0)
|
|
ctx->result = buffer->status;
|
|
}
|
|
|
|
static void write_sm4_signature(struct hlsl_ctx *ctx, struct dxbc_writer *dxbc,
|
|
const struct hlsl_ir_function_decl *func, bool output)
|
|
{
|
|
struct vkd3d_bytecode_buffer buffer = {0};
|
|
struct vkd3d_string_buffer *string;
|
|
const struct hlsl_ir_var *var;
|
|
size_t count_position;
|
|
unsigned int i;
|
|
bool ret;
|
|
|
|
count_position = put_u32(&buffer, 0);
|
|
put_u32(&buffer, 8); /* unknown */
|
|
|
|
LIST_FOR_EACH_ENTRY(var, &func->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
unsigned int width = (1u << var->data_type->dimx) - 1, use_mask;
|
|
enum vkd3d_shader_sysval_semantic semantic;
|
|
uint32_t usage_idx, reg_idx;
|
|
bool has_idx;
|
|
|
|
if ((output && !var->is_output_semantic) || (!output && !var->is_input_semantic))
|
|
continue;
|
|
|
|
ret = sysval_semantic_from_hlsl(&semantic, ctx, &var->semantic, output);
|
|
VKD3D_ASSERT(ret);
|
|
if (semantic == ~0u)
|
|
continue;
|
|
usage_idx = var->semantic.index;
|
|
|
|
if (hlsl_sm4_register_from_semantic(ctx, &var->semantic, output, NULL, &has_idx))
|
|
{
|
|
reg_idx = has_idx ? var->semantic.index : ~0u;
|
|
}
|
|
else
|
|
{
|
|
VKD3D_ASSERT(var->regs[HLSL_REGSET_NUMERIC].allocated);
|
|
reg_idx = var->regs[HLSL_REGSET_NUMERIC].id;
|
|
}
|
|
|
|
use_mask = width; /* FIXME: accurately report use mask */
|
|
if (output)
|
|
use_mask = 0xf ^ use_mask;
|
|
|
|
/* Special pixel shader semantics (TARGET, DEPTH, COVERAGE). */
|
|
if (semantic >= VKD3D_SHADER_SV_TARGET)
|
|
semantic = VKD3D_SHADER_SV_NONE;
|
|
|
|
put_u32(&buffer, 0); /* name */
|
|
put_u32(&buffer, usage_idx);
|
|
put_u32(&buffer, semantic);
|
|
switch (var->data_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
case HLSL_TYPE_HALF:
|
|
put_u32(&buffer, VKD3D_SHADER_COMPONENT_FLOAT);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
put_u32(&buffer, VKD3D_SHADER_COMPONENT_INT);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_UINT:
|
|
put_u32(&buffer, VKD3D_SHADER_COMPONENT_UINT);
|
|
break;
|
|
|
|
default:
|
|
if ((string = hlsl_type_to_string(ctx, var->data_type)))
|
|
hlsl_error(ctx, &var->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TYPE,
|
|
"Invalid data type %s for semantic variable %s.", string->buffer, var->name);
|
|
hlsl_release_string_buffer(ctx, string);
|
|
put_u32(&buffer, VKD3D_SHADER_COMPONENT_VOID);
|
|
}
|
|
put_u32(&buffer, reg_idx);
|
|
put_u32(&buffer, vkd3d_make_u16(width, use_mask));
|
|
}
|
|
|
|
i = 0;
|
|
LIST_FOR_EACH_ENTRY(var, &func->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
enum vkd3d_shader_sysval_semantic semantic;
|
|
const char *name = var->semantic.name;
|
|
size_t string_offset;
|
|
|
|
if ((output && !var->is_output_semantic) || (!output && !var->is_input_semantic))
|
|
continue;
|
|
|
|
sysval_semantic_from_hlsl(&semantic, ctx, &var->semantic, output);
|
|
if (semantic == ~0u)
|
|
continue;
|
|
|
|
if (semantic == VKD3D_SHADER_SV_TARGET && !ascii_strcasecmp(name, "color"))
|
|
string_offset = put_string(&buffer, "SV_Target");
|
|
else if (semantic == VKD3D_SHADER_SV_DEPTH && !ascii_strcasecmp(name, "depth"))
|
|
string_offset = put_string(&buffer, "SV_Depth");
|
|
else if (semantic == VKD3D_SHADER_SV_POSITION && !ascii_strcasecmp(name, "position"))
|
|
string_offset = put_string(&buffer, "SV_Position");
|
|
else
|
|
string_offset = put_string(&buffer, name);
|
|
set_u32(&buffer, (2 + i++ * 6) * sizeof(uint32_t), string_offset);
|
|
}
|
|
|
|
set_u32(&buffer, count_position, i);
|
|
|
|
add_section(ctx, dxbc, output ? TAG_OSGN : TAG_ISGN, &buffer);
|
|
}
|
|
|
|
static D3D_SHADER_VARIABLE_CLASS sm4_class(const struct hlsl_type *type)
|
|
{
|
|
switch (type->class)
|
|
{
|
|
case HLSL_CLASS_MATRIX:
|
|
VKD3D_ASSERT(type->modifiers & HLSL_MODIFIERS_MAJORITY_MASK);
|
|
if (type->modifiers & HLSL_MODIFIER_COLUMN_MAJOR)
|
|
return D3D_SVC_MATRIX_COLUMNS;
|
|
else
|
|
return D3D_SVC_MATRIX_ROWS;
|
|
case HLSL_CLASS_SCALAR:
|
|
return D3D_SVC_SCALAR;
|
|
case HLSL_CLASS_VECTOR:
|
|
return D3D_SVC_VECTOR;
|
|
|
|
case HLSL_CLASS_ARRAY:
|
|
case HLSL_CLASS_DEPTH_STENCIL_STATE:
|
|
case HLSL_CLASS_DEPTH_STENCIL_VIEW:
|
|
case HLSL_CLASS_EFFECT_GROUP:
|
|
case HLSL_CLASS_ERROR:
|
|
case HLSL_CLASS_STRUCT:
|
|
case HLSL_CLASS_PASS:
|
|
case HLSL_CLASS_PIXEL_SHADER:
|
|
case HLSL_CLASS_RASTERIZER_STATE:
|
|
case HLSL_CLASS_RENDER_TARGET_VIEW:
|
|
case HLSL_CLASS_SAMPLER:
|
|
case HLSL_CLASS_STRING:
|
|
case HLSL_CLASS_TECHNIQUE:
|
|
case HLSL_CLASS_TEXTURE:
|
|
case HLSL_CLASS_UAV:
|
|
case HLSL_CLASS_VERTEX_SHADER:
|
|
case HLSL_CLASS_VOID:
|
|
case HLSL_CLASS_CONSTANT_BUFFER:
|
|
case HLSL_CLASS_COMPUTE_SHADER:
|
|
case HLSL_CLASS_DOMAIN_SHADER:
|
|
case HLSL_CLASS_HULL_SHADER:
|
|
case HLSL_CLASS_GEOMETRY_SHADER:
|
|
case HLSL_CLASS_BLEND_STATE:
|
|
case HLSL_CLASS_NULL:
|
|
break;
|
|
}
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
static D3D_SHADER_VARIABLE_TYPE sm4_base_type(const struct hlsl_type *type)
|
|
{
|
|
switch (type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_BOOL:
|
|
return D3D_SVT_BOOL;
|
|
case HLSL_TYPE_DOUBLE:
|
|
return D3D_SVT_DOUBLE;
|
|
case HLSL_TYPE_FLOAT:
|
|
case HLSL_TYPE_HALF:
|
|
return D3D_SVT_FLOAT;
|
|
case HLSL_TYPE_INT:
|
|
return D3D_SVT_INT;
|
|
case HLSL_TYPE_UINT:
|
|
return D3D_SVT_UINT;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void write_sm4_type(struct hlsl_ctx *ctx, struct vkd3d_bytecode_buffer *buffer, struct hlsl_type *type)
|
|
{
|
|
const struct hlsl_type *array_type = hlsl_get_multiarray_element_type(type);
|
|
const char *name = array_type->name ? array_type->name : "<unnamed>";
|
|
const struct hlsl_profile_info *profile = ctx->profile;
|
|
unsigned int array_size = 0;
|
|
size_t name_offset = 0;
|
|
size_t i;
|
|
|
|
if (type->bytecode_offset)
|
|
return;
|
|
|
|
if (profile->major_version >= 5)
|
|
name_offset = put_string(buffer, name);
|
|
|
|
if (type->class == HLSL_CLASS_ARRAY)
|
|
array_size = hlsl_get_multiarray_size(type);
|
|
|
|
if (array_type->class == HLSL_CLASS_STRUCT)
|
|
{
|
|
unsigned int field_count = 0;
|
|
size_t fields_offset = 0;
|
|
|
|
for (i = 0; i < array_type->e.record.field_count; ++i)
|
|
{
|
|
struct hlsl_struct_field *field = &array_type->e.record.fields[i];
|
|
|
|
if (!field->type->reg_size[HLSL_REGSET_NUMERIC])
|
|
continue;
|
|
|
|
field->name_bytecode_offset = put_string(buffer, field->name);
|
|
write_sm4_type(ctx, buffer, field->type);
|
|
++field_count;
|
|
}
|
|
|
|
fields_offset = bytecode_align(buffer);
|
|
|
|
for (i = 0; i < array_type->e.record.field_count; ++i)
|
|
{
|
|
struct hlsl_struct_field *field = &array_type->e.record.fields[i];
|
|
|
|
if (!field->type->reg_size[HLSL_REGSET_NUMERIC])
|
|
continue;
|
|
|
|
put_u32(buffer, field->name_bytecode_offset);
|
|
put_u32(buffer, field->type->bytecode_offset);
|
|
put_u32(buffer, field->reg_offset[HLSL_REGSET_NUMERIC] * sizeof(float));
|
|
}
|
|
type->bytecode_offset = put_u32(buffer, vkd3d_make_u32(D3D_SVC_STRUCT, D3D_SVT_VOID));
|
|
put_u32(buffer, vkd3d_make_u32(1, hlsl_type_component_count(array_type)));
|
|
put_u32(buffer, vkd3d_make_u32(array_size, field_count));
|
|
put_u32(buffer, fields_offset);
|
|
}
|
|
else
|
|
{
|
|
VKD3D_ASSERT(array_type->class <= HLSL_CLASS_LAST_NUMERIC);
|
|
type->bytecode_offset = put_u32(buffer, vkd3d_make_u32(sm4_class(array_type), sm4_base_type(array_type)));
|
|
put_u32(buffer, vkd3d_make_u32(array_type->dimy, array_type->dimx));
|
|
put_u32(buffer, vkd3d_make_u32(array_size, 0));
|
|
put_u32(buffer, 1);
|
|
}
|
|
|
|
if (profile->major_version >= 5)
|
|
{
|
|
put_u32(buffer, 0); /* FIXME: unknown */
|
|
put_u32(buffer, 0); /* FIXME: unknown */
|
|
put_u32(buffer, 0); /* FIXME: unknown */
|
|
put_u32(buffer, 0); /* FIXME: unknown */
|
|
put_u32(buffer, name_offset);
|
|
}
|
|
}
|
|
|
|
static D3D_SHADER_INPUT_TYPE sm4_resource_type(const struct hlsl_type *type)
|
|
{
|
|
switch (type->class)
|
|
{
|
|
case HLSL_CLASS_SAMPLER:
|
|
return D3D_SIT_SAMPLER;
|
|
case HLSL_CLASS_TEXTURE:
|
|
return D3D_SIT_TEXTURE;
|
|
case HLSL_CLASS_UAV:
|
|
return D3D_SIT_UAV_RWTYPED;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
static enum vkd3d_sm4_data_type sm4_data_type(const struct hlsl_type *type)
|
|
{
|
|
switch (type->e.resource.format->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_DOUBLE:
|
|
return VKD3D_SM4_DATA_DOUBLE;
|
|
|
|
case HLSL_TYPE_FLOAT:
|
|
case HLSL_TYPE_HALF:
|
|
return VKD3D_SM4_DATA_FLOAT;
|
|
|
|
case HLSL_TYPE_INT:
|
|
return VKD3D_SM4_DATA_INT;
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_UINT:
|
|
return VKD3D_SM4_DATA_UINT;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static D3D_SRV_DIMENSION sm4_rdef_resource_dimension(const struct hlsl_type *type)
|
|
{
|
|
switch (type->sampler_dim)
|
|
{
|
|
case HLSL_SAMPLER_DIM_1D:
|
|
return D3D_SRV_DIMENSION_TEXTURE1D;
|
|
case HLSL_SAMPLER_DIM_2D:
|
|
return D3D_SRV_DIMENSION_TEXTURE2D;
|
|
case HLSL_SAMPLER_DIM_3D:
|
|
return D3D_SRV_DIMENSION_TEXTURE3D;
|
|
case HLSL_SAMPLER_DIM_CUBE:
|
|
return D3D_SRV_DIMENSION_TEXTURECUBE;
|
|
case HLSL_SAMPLER_DIM_1DARRAY:
|
|
return D3D_SRV_DIMENSION_TEXTURE1DARRAY;
|
|
case HLSL_SAMPLER_DIM_2DARRAY:
|
|
return D3D_SRV_DIMENSION_TEXTURE2DARRAY;
|
|
case HLSL_SAMPLER_DIM_2DMS:
|
|
return D3D_SRV_DIMENSION_TEXTURE2DMS;
|
|
case HLSL_SAMPLER_DIM_2DMSARRAY:
|
|
return D3D_SRV_DIMENSION_TEXTURE2DMSARRAY;
|
|
case HLSL_SAMPLER_DIM_CUBEARRAY:
|
|
return D3D_SRV_DIMENSION_TEXTURECUBEARRAY;
|
|
case HLSL_SAMPLER_DIM_BUFFER:
|
|
case HLSL_SAMPLER_DIM_STRUCTURED_BUFFER:
|
|
return D3D_SRV_DIMENSION_BUFFER;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
struct extern_resource
|
|
{
|
|
/* var is only not NULL if this resource is a whole variable, so it may be responsible for more
|
|
* than one component. */
|
|
const struct hlsl_ir_var *var;
|
|
const struct hlsl_buffer *buffer;
|
|
|
|
char *name;
|
|
bool is_user_packed;
|
|
|
|
/* The data type of a single component of the resource.
|
|
* This might be different from the data type of the resource itself in 4.0
|
|
* profiles, where an array (or multi-dimensional array) is handled as a
|
|
* single resource, unlike in 5.0. */
|
|
struct hlsl_type *component_type;
|
|
|
|
enum hlsl_regset regset;
|
|
unsigned int id, space, index, bind_count;
|
|
|
|
struct vkd3d_shader_location loc;
|
|
};
|
|
|
|
static int sm4_compare_extern_resources(const void *a, const void *b)
|
|
{
|
|
const struct extern_resource *aa = (const struct extern_resource *)a;
|
|
const struct extern_resource *bb = (const struct extern_resource *)b;
|
|
int r;
|
|
|
|
if ((r = vkd3d_u32_compare(aa->regset, bb->regset)))
|
|
return r;
|
|
|
|
if ((r = vkd3d_u32_compare(aa->space, bb->space)))
|
|
return r;
|
|
|
|
return vkd3d_u32_compare(aa->index, bb->index);
|
|
}
|
|
|
|
static void sm4_free_extern_resources(struct extern_resource *extern_resources, unsigned int count)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < count; ++i)
|
|
vkd3d_free(extern_resources[i].name);
|
|
vkd3d_free(extern_resources);
|
|
}
|
|
|
|
static const char *string_skip_tag(const char *string)
|
|
{
|
|
if (!strncmp(string, "<resource>", strlen("<resource>")))
|
|
return string + strlen("<resource>");
|
|
return string;
|
|
}
|
|
|
|
static struct extern_resource *sm4_get_extern_resources(struct hlsl_ctx *ctx, unsigned int *count)
|
|
{
|
|
bool separate_components = ctx->profile->major_version == 5 && ctx->profile->minor_version == 0;
|
|
struct extern_resource *extern_resources = NULL;
|
|
const struct hlsl_ir_var *var;
|
|
struct hlsl_buffer *buffer;
|
|
enum hlsl_regset regset;
|
|
size_t capacity = 0;
|
|
char *name;
|
|
|
|
*count = 0;
|
|
|
|
LIST_FOR_EACH_ENTRY(var, &ctx->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
if (separate_components)
|
|
{
|
|
unsigned int component_count = hlsl_type_component_count(var->data_type);
|
|
unsigned int k, regset_offset;
|
|
|
|
for (k = 0; k < component_count; ++k)
|
|
{
|
|
struct hlsl_type *component_type = hlsl_type_get_component_type(ctx, var->data_type, k);
|
|
struct vkd3d_string_buffer *name_buffer;
|
|
|
|
if (!hlsl_type_is_resource(component_type))
|
|
continue;
|
|
|
|
regset_offset = hlsl_type_get_component_offset(ctx, var->data_type, k, ®set);
|
|
|
|
if (regset_offset > var->regs[regset].allocation_size)
|
|
continue;
|
|
|
|
if (var->objects_usage[regset][regset_offset].used)
|
|
{
|
|
if (!(hlsl_array_reserve(ctx, (void **)&extern_resources, &capacity, *count + 1,
|
|
sizeof(*extern_resources))))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
|
|
if (!(name_buffer = hlsl_component_to_string(ctx, var, k)))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
if (!(name = hlsl_strdup(ctx, string_skip_tag(name_buffer->buffer))))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
hlsl_release_string_buffer(ctx, name_buffer);
|
|
return NULL;
|
|
}
|
|
hlsl_release_string_buffer(ctx, name_buffer);
|
|
|
|
extern_resources[*count].var = NULL;
|
|
extern_resources[*count].buffer = NULL;
|
|
|
|
extern_resources[*count].name = name;
|
|
extern_resources[*count].is_user_packed = !!var->reg_reservation.reg_type;
|
|
|
|
extern_resources[*count].component_type = component_type;
|
|
|
|
extern_resources[*count].regset = regset;
|
|
extern_resources[*count].id = var->regs[regset].id;
|
|
extern_resources[*count].space = var->regs[regset].space;
|
|
extern_resources[*count].index = var->regs[regset].index + regset_offset;
|
|
extern_resources[*count].bind_count = 1;
|
|
extern_resources[*count].loc = var->loc;
|
|
|
|
++*count;
|
|
}
|
|
}
|
|
}
|
|
else
|
|
{
|
|
unsigned int r;
|
|
|
|
if (!hlsl_type_is_resource(var->data_type))
|
|
continue;
|
|
|
|
for (r = 0; r <= HLSL_REGSET_LAST; ++r)
|
|
{
|
|
if (!var->regs[r].allocated)
|
|
continue;
|
|
|
|
if (!(hlsl_array_reserve(ctx, (void **)&extern_resources, &capacity, *count + 1,
|
|
sizeof(*extern_resources))))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
|
|
if (!(name = hlsl_strdup(ctx, string_skip_tag(var->name))))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
|
|
extern_resources[*count].var = var;
|
|
extern_resources[*count].buffer = NULL;
|
|
|
|
extern_resources[*count].name = name;
|
|
/* For some reason 5.1 resources aren't marked as
|
|
* user-packed, but cbuffers still are. */
|
|
extern_resources[*count].is_user_packed = hlsl_version_lt(ctx, 5, 1)
|
|
&& !!var->reg_reservation.reg_type;
|
|
|
|
extern_resources[*count].component_type = hlsl_type_get_component_type(ctx, var->data_type, 0);
|
|
|
|
extern_resources[*count].regset = r;
|
|
extern_resources[*count].id = var->regs[r].id;
|
|
extern_resources[*count].space = var->regs[r].space;
|
|
extern_resources[*count].index = var->regs[r].index;
|
|
extern_resources[*count].bind_count = var->bind_count[r];
|
|
extern_resources[*count].loc = var->loc;
|
|
|
|
++*count;
|
|
}
|
|
}
|
|
}
|
|
|
|
LIST_FOR_EACH_ENTRY(buffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
if (!buffer->reg.allocated)
|
|
continue;
|
|
|
|
if (!(hlsl_array_reserve(ctx, (void **)&extern_resources, &capacity, *count + 1,
|
|
sizeof(*extern_resources))))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
|
|
if (!(name = hlsl_strdup(ctx, buffer->name)))
|
|
{
|
|
sm4_free_extern_resources(extern_resources, *count);
|
|
*count = 0;
|
|
return NULL;
|
|
}
|
|
|
|
extern_resources[*count].var = NULL;
|
|
extern_resources[*count].buffer = buffer;
|
|
|
|
extern_resources[*count].name = name;
|
|
extern_resources[*count].is_user_packed = !!buffer->reservation.reg_type;
|
|
|
|
extern_resources[*count].component_type = NULL;
|
|
|
|
extern_resources[*count].regset = HLSL_REGSET_NUMERIC;
|
|
extern_resources[*count].id = buffer->reg.id;
|
|
extern_resources[*count].space = buffer->reg.space;
|
|
extern_resources[*count].index = buffer->reg.index;
|
|
extern_resources[*count].bind_count = 1;
|
|
extern_resources[*count].loc = buffer->loc;
|
|
|
|
++*count;
|
|
}
|
|
|
|
qsort(extern_resources, *count, sizeof(*extern_resources), sm4_compare_extern_resources);
|
|
return extern_resources;
|
|
}
|
|
|
|
/* For some reason, for matrices, values from default value initializers end up in different
|
|
* components than from regular initializers. Default value initializers fill the matrix in
|
|
* vertical reading order (left-to-right top-to-bottom) instead of regular reading order
|
|
* (top-to-bottom left-to-right), so they have to be adjusted.
|
|
* An exception is that the order of matrix initializers for function parameters are row-major
|
|
* (top-to-bottom left-to-right). */
|
|
static unsigned int get_component_index_from_default_initializer_index(struct hlsl_type *type, unsigned int index)
|
|
{
|
|
unsigned int element_comp_count, element, x, y, i;
|
|
unsigned int base = 0;
|
|
|
|
switch (type->class)
|
|
{
|
|
case HLSL_CLASS_MATRIX:
|
|
x = index / type->dimy;
|
|
y = index % type->dimy;
|
|
return y * type->dimx + x;
|
|
|
|
case HLSL_CLASS_ARRAY:
|
|
element_comp_count = hlsl_type_component_count(type->e.array.type);
|
|
element = index / element_comp_count;
|
|
base = element * element_comp_count;
|
|
return base + get_component_index_from_default_initializer_index(type->e.array.type, index - base);
|
|
|
|
case HLSL_CLASS_STRUCT:
|
|
for (i = 0; i < type->e.record.field_count; ++i)
|
|
{
|
|
struct hlsl_type *field_type = type->e.record.fields[i].type;
|
|
|
|
element_comp_count = hlsl_type_component_count(field_type);
|
|
if (index - base < element_comp_count)
|
|
return base + get_component_index_from_default_initializer_index(field_type, index - base);
|
|
base += element_comp_count;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
return index;
|
|
}
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
static void write_sm4_rdef(struct hlsl_ctx *ctx, struct dxbc_writer *dxbc)
|
|
{
|
|
uint32_t binding_desc_size = (hlsl_version_ge(ctx, 5, 1) ? 10 : 8) * sizeof(uint32_t);
|
|
size_t cbuffers_offset, resources_offset, creator_offset, string_offset;
|
|
unsigned int cbuffer_count = 0, extern_resources_count, i, j;
|
|
size_t cbuffer_position, resource_position, creator_position;
|
|
const struct hlsl_profile_info *profile = ctx->profile;
|
|
struct vkd3d_bytecode_buffer buffer = {0};
|
|
struct extern_resource *extern_resources;
|
|
const struct hlsl_buffer *cbuffer;
|
|
const struct hlsl_ir_var *var;
|
|
|
|
static const uint16_t target_types[] =
|
|
{
|
|
0xffff, /* PIXEL */
|
|
0xfffe, /* VERTEX */
|
|
0x4753, /* GEOMETRY */
|
|
0x4853, /* HULL */
|
|
0x4453, /* DOMAIN */
|
|
0x4353, /* COMPUTE */
|
|
};
|
|
|
|
extern_resources = sm4_get_extern_resources(ctx, &extern_resources_count);
|
|
|
|
LIST_FOR_EACH_ENTRY(cbuffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
if (cbuffer->reg.allocated)
|
|
++cbuffer_count;
|
|
}
|
|
|
|
put_u32(&buffer, cbuffer_count);
|
|
cbuffer_position = put_u32(&buffer, 0);
|
|
put_u32(&buffer, extern_resources_count);
|
|
resource_position = put_u32(&buffer, 0);
|
|
put_u32(&buffer, vkd3d_make_u32(vkd3d_make_u16(profile->minor_version, profile->major_version),
|
|
target_types[profile->type]));
|
|
put_u32(&buffer, 0); /* FIXME: compilation flags */
|
|
creator_position = put_u32(&buffer, 0);
|
|
|
|
if (profile->major_version >= 5)
|
|
{
|
|
put_u32(&buffer, hlsl_version_ge(ctx, 5, 1) ? TAG_RD11_REVERSE : TAG_RD11);
|
|
put_u32(&buffer, 15 * sizeof(uint32_t)); /* size of RDEF header including this header */
|
|
put_u32(&buffer, 6 * sizeof(uint32_t)); /* size of buffer desc */
|
|
put_u32(&buffer, binding_desc_size); /* size of binding desc */
|
|
put_u32(&buffer, 10 * sizeof(uint32_t)); /* size of variable desc */
|
|
put_u32(&buffer, 9 * sizeof(uint32_t)); /* size of type desc */
|
|
put_u32(&buffer, 3 * sizeof(uint32_t)); /* size of member desc */
|
|
put_u32(&buffer, 0); /* unknown; possibly a null terminator */
|
|
}
|
|
|
|
/* Bound resources. */
|
|
|
|
resources_offset = bytecode_align(&buffer);
|
|
set_u32(&buffer, resource_position, resources_offset);
|
|
|
|
for (i = 0; i < extern_resources_count; ++i)
|
|
{
|
|
const struct extern_resource *resource = &extern_resources[i];
|
|
uint32_t flags = 0;
|
|
|
|
if (resource->is_user_packed)
|
|
flags |= D3D_SIF_USERPACKED;
|
|
|
|
put_u32(&buffer, 0); /* name */
|
|
if (resource->buffer)
|
|
put_u32(&buffer, resource->buffer->type == HLSL_BUFFER_CONSTANT ? D3D_SIT_CBUFFER : D3D_SIT_TBUFFER);
|
|
else
|
|
put_u32(&buffer, sm4_resource_type(resource->component_type));
|
|
if (resource->regset == HLSL_REGSET_TEXTURES || resource->regset == HLSL_REGSET_UAVS)
|
|
{
|
|
unsigned int dimx = resource->component_type->e.resource.format->dimx;
|
|
|
|
put_u32(&buffer, sm4_data_type(resource->component_type));
|
|
put_u32(&buffer, sm4_rdef_resource_dimension(resource->component_type));
|
|
put_u32(&buffer, ~0u); /* FIXME: multisample count */
|
|
flags |= (dimx - 1) << VKD3D_SM4_SIF_TEXTURE_COMPONENTS_SHIFT;
|
|
}
|
|
else
|
|
{
|
|
put_u32(&buffer, 0);
|
|
put_u32(&buffer, 0);
|
|
put_u32(&buffer, 0);
|
|
}
|
|
put_u32(&buffer, resource->index);
|
|
put_u32(&buffer, resource->bind_count);
|
|
put_u32(&buffer, flags);
|
|
|
|
if (hlsl_version_ge(ctx, 5, 1))
|
|
{
|
|
put_u32(&buffer, resource->space);
|
|
put_u32(&buffer, resource->id);
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < extern_resources_count; ++i)
|
|
{
|
|
const struct extern_resource *resource = &extern_resources[i];
|
|
|
|
string_offset = put_string(&buffer, resource->name);
|
|
set_u32(&buffer, resources_offset + i * binding_desc_size, string_offset);
|
|
}
|
|
|
|
/* Buffers. */
|
|
|
|
cbuffers_offset = bytecode_align(&buffer);
|
|
set_u32(&buffer, cbuffer_position, cbuffers_offset);
|
|
LIST_FOR_EACH_ENTRY(cbuffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
unsigned int var_count = 0;
|
|
|
|
if (!cbuffer->reg.allocated)
|
|
continue;
|
|
|
|
LIST_FOR_EACH_ENTRY(var, &ctx->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
if (var->is_uniform && var->buffer == cbuffer && var->data_type->reg_size[HLSL_REGSET_NUMERIC])
|
|
++var_count;
|
|
}
|
|
|
|
put_u32(&buffer, 0); /* name */
|
|
put_u32(&buffer, var_count);
|
|
put_u32(&buffer, 0); /* variable offset */
|
|
put_u32(&buffer, align(cbuffer->size, 4) * sizeof(float));
|
|
put_u32(&buffer, 0); /* FIXME: flags */
|
|
put_u32(&buffer, cbuffer->type == HLSL_BUFFER_CONSTANT ? D3D_CT_CBUFFER : D3D_CT_TBUFFER);
|
|
}
|
|
|
|
i = 0;
|
|
LIST_FOR_EACH_ENTRY(cbuffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
if (!cbuffer->reg.allocated)
|
|
continue;
|
|
|
|
string_offset = put_string(&buffer, cbuffer->name);
|
|
set_u32(&buffer, cbuffers_offset + i++ * 6 * sizeof(uint32_t), string_offset);
|
|
}
|
|
|
|
i = 0;
|
|
LIST_FOR_EACH_ENTRY(cbuffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
size_t vars_start = bytecode_align(&buffer);
|
|
|
|
if (!cbuffer->reg.allocated)
|
|
continue;
|
|
|
|
set_u32(&buffer, cbuffers_offset + (i++ * 6 + 2) * sizeof(uint32_t), vars_start);
|
|
|
|
LIST_FOR_EACH_ENTRY(var, &ctx->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
if (var->is_uniform && var->buffer == cbuffer && var->data_type->reg_size[HLSL_REGSET_NUMERIC])
|
|
{
|
|
uint32_t flags = 0;
|
|
|
|
if (var->is_read)
|
|
flags |= D3D_SVF_USED;
|
|
|
|
put_u32(&buffer, 0); /* name */
|
|
put_u32(&buffer, var->buffer_offset * sizeof(float));
|
|
put_u32(&buffer, var->data_type->reg_size[HLSL_REGSET_NUMERIC] * sizeof(float));
|
|
put_u32(&buffer, flags);
|
|
put_u32(&buffer, 0); /* type */
|
|
put_u32(&buffer, 0); /* default value */
|
|
|
|
if (profile->major_version >= 5)
|
|
{
|
|
put_u32(&buffer, 0); /* texture start */
|
|
put_u32(&buffer, 0); /* texture count */
|
|
put_u32(&buffer, 0); /* sampler start */
|
|
put_u32(&buffer, 0); /* sampler count */
|
|
}
|
|
}
|
|
}
|
|
|
|
j = 0;
|
|
LIST_FOR_EACH_ENTRY(var, &ctx->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
if (var->is_uniform && var->buffer == cbuffer && var->data_type->reg_size[HLSL_REGSET_NUMERIC])
|
|
{
|
|
const unsigned int var_size = (profile->major_version >= 5 ? 10 : 6);
|
|
size_t var_offset = vars_start + j * var_size * sizeof(uint32_t);
|
|
|
|
string_offset = put_string(&buffer, var->name);
|
|
set_u32(&buffer, var_offset, string_offset);
|
|
write_sm4_type(ctx, &buffer, var->data_type);
|
|
set_u32(&buffer, var_offset + 4 * sizeof(uint32_t), var->data_type->bytecode_offset);
|
|
|
|
if (var->default_values)
|
|
{
|
|
unsigned int reg_size = var->data_type->reg_size[HLSL_REGSET_NUMERIC];
|
|
unsigned int comp_count = hlsl_type_component_count(var->data_type);
|
|
unsigned int default_value_offset;
|
|
unsigned int k;
|
|
|
|
default_value_offset = bytecode_reserve_bytes(&buffer, reg_size * sizeof(uint32_t));
|
|
set_u32(&buffer, var_offset + 5 * sizeof(uint32_t), default_value_offset);
|
|
|
|
for (k = 0; k < comp_count; ++k)
|
|
{
|
|
struct hlsl_type *comp_type = hlsl_type_get_component_type(ctx, var->data_type, k);
|
|
unsigned int comp_offset, comp_index;
|
|
enum hlsl_regset regset;
|
|
|
|
if (comp_type->class == HLSL_CLASS_STRING)
|
|
{
|
|
hlsl_error(ctx, &var->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TYPE,
|
|
"Cannot write string default value.");
|
|
continue;
|
|
}
|
|
|
|
comp_index = get_component_index_from_default_initializer_index(var->data_type, k);
|
|
comp_offset = hlsl_type_get_component_offset(ctx, var->data_type, comp_index, ®set);
|
|
if (regset == HLSL_REGSET_NUMERIC)
|
|
{
|
|
if (comp_type->e.numeric.type == HLSL_TYPE_DOUBLE)
|
|
hlsl_fixme(ctx, &var->loc, "Write double default values.");
|
|
|
|
set_u32(&buffer, default_value_offset + comp_offset * sizeof(uint32_t),
|
|
var->default_values[k].number.u);
|
|
}
|
|
}
|
|
}
|
|
++j;
|
|
}
|
|
}
|
|
}
|
|
|
|
creator_offset = put_string(&buffer, vkd3d_shader_get_version(NULL, NULL));
|
|
set_u32(&buffer, creator_position, creator_offset);
|
|
|
|
add_section(ctx, dxbc, TAG_RDEF, &buffer);
|
|
|
|
sm4_free_extern_resources(extern_resources, extern_resources_count);
|
|
}
|
|
|
|
static enum vkd3d_sm4_resource_type sm4_resource_dimension(const struct hlsl_type *type)
|
|
{
|
|
switch (type->sampler_dim)
|
|
{
|
|
case HLSL_SAMPLER_DIM_1D:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_1D;
|
|
case HLSL_SAMPLER_DIM_2D:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_2D;
|
|
case HLSL_SAMPLER_DIM_3D:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_3D;
|
|
case HLSL_SAMPLER_DIM_CUBE:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_CUBE;
|
|
case HLSL_SAMPLER_DIM_1DARRAY:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_1DARRAY;
|
|
case HLSL_SAMPLER_DIM_2DARRAY:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_2DARRAY;
|
|
case HLSL_SAMPLER_DIM_2DMS:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_2DMS;
|
|
case HLSL_SAMPLER_DIM_2DMSARRAY:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_2DMSARRAY;
|
|
case HLSL_SAMPLER_DIM_CUBEARRAY:
|
|
return VKD3D_SM4_RESOURCE_TEXTURE_CUBEARRAY;
|
|
case HLSL_SAMPLER_DIM_BUFFER:
|
|
case HLSL_SAMPLER_DIM_STRUCTURED_BUFFER:
|
|
return VKD3D_SM4_RESOURCE_BUFFER;
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
struct sm4_instruction_modifier
|
|
{
|
|
enum vkd3d_sm4_instruction_modifier type;
|
|
|
|
union
|
|
{
|
|
struct
|
|
{
|
|
int u, v, w;
|
|
} aoffimmi;
|
|
} u;
|
|
};
|
|
|
|
static uint32_t sm4_encode_instruction_modifier(const struct sm4_instruction_modifier *imod)
|
|
{
|
|
uint32_t word = 0;
|
|
|
|
word |= VKD3D_SM4_MODIFIER_MASK & imod->type;
|
|
|
|
switch (imod->type)
|
|
{
|
|
case VKD3D_SM4_MODIFIER_AOFFIMMI:
|
|
VKD3D_ASSERT(-8 <= imod->u.aoffimmi.u && imod->u.aoffimmi.u <= 7);
|
|
VKD3D_ASSERT(-8 <= imod->u.aoffimmi.v && imod->u.aoffimmi.v <= 7);
|
|
VKD3D_ASSERT(-8 <= imod->u.aoffimmi.w && imod->u.aoffimmi.w <= 7);
|
|
word |= ((uint32_t)imod->u.aoffimmi.u & 0xf) << VKD3D_SM4_AOFFIMMI_U_SHIFT;
|
|
word |= ((uint32_t)imod->u.aoffimmi.v & 0xf) << VKD3D_SM4_AOFFIMMI_V_SHIFT;
|
|
word |= ((uint32_t)imod->u.aoffimmi.w & 0xf) << VKD3D_SM4_AOFFIMMI_W_SHIFT;
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
return word;
|
|
}
|
|
|
|
struct sm4_instruction
|
|
{
|
|
enum vkd3d_sm4_opcode opcode;
|
|
uint32_t extra_bits;
|
|
|
|
struct sm4_instruction_modifier modifiers[1];
|
|
unsigned int modifier_count;
|
|
|
|
struct vkd3d_shader_dst_param dsts[2];
|
|
unsigned int dst_count;
|
|
|
|
struct vkd3d_shader_src_param srcs[5];
|
|
unsigned int src_count;
|
|
|
|
unsigned int byte_stride;
|
|
|
|
uint32_t idx[3];
|
|
unsigned int idx_count;
|
|
|
|
struct vkd3d_shader_src_param idx_srcs[7];
|
|
unsigned int idx_src_count;
|
|
};
|
|
|
|
static void sm4_register_from_node(struct vkd3d_shader_register *reg, uint32_t *writemask,
|
|
const struct hlsl_ir_node *instr)
|
|
{
|
|
VKD3D_ASSERT(instr->reg.allocated);
|
|
reg->type = VKD3DSPR_TEMP;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
reg->idx[0].offset = instr->reg.id;
|
|
reg->idx_count = 1;
|
|
*writemask = instr->reg.writemask;
|
|
}
|
|
|
|
static void sm4_numeric_register_from_deref(struct hlsl_ctx *ctx, struct vkd3d_shader_register *reg,
|
|
enum vkd3d_shader_register_type type, uint32_t *writemask, const struct hlsl_deref *deref,
|
|
struct sm4_instruction *sm4_instr)
|
|
{
|
|
const struct hlsl_ir_var *var = deref->var;
|
|
unsigned int offset_const_deref;
|
|
|
|
reg->type = type;
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_NUMERIC].id;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
|
|
VKD3D_ASSERT(var->regs[HLSL_REGSET_NUMERIC].allocated);
|
|
|
|
if (!var->indexable)
|
|
{
|
|
offset_const_deref = hlsl_offset_from_deref_safe(ctx, deref);
|
|
reg->idx[0].offset += offset_const_deref / 4;
|
|
reg->idx_count = 1;
|
|
}
|
|
else
|
|
{
|
|
offset_const_deref = deref->const_offset;
|
|
reg->idx[1].offset = offset_const_deref / 4;
|
|
reg->idx_count = 2;
|
|
|
|
if (deref->rel_offset.node)
|
|
{
|
|
struct vkd3d_shader_src_param *idx_src;
|
|
unsigned int idx_writemask;
|
|
|
|
VKD3D_ASSERT(sm4_instr->idx_src_count < ARRAY_SIZE(sm4_instr->idx_srcs));
|
|
idx_src = &sm4_instr->idx_srcs[sm4_instr->idx_src_count++];
|
|
memset(idx_src, 0, sizeof(*idx_src));
|
|
|
|
reg->idx[1].rel_addr = idx_src;
|
|
sm4_register_from_node(&idx_src->reg, &idx_writemask, deref->rel_offset.node);
|
|
VKD3D_ASSERT(idx_writemask != 0);
|
|
idx_src->swizzle = swizzle_from_sm4(hlsl_swizzle_from_writemask(idx_writemask));
|
|
}
|
|
}
|
|
|
|
*writemask = 0xf & (0xf << (offset_const_deref % 4));
|
|
if (var->regs[HLSL_REGSET_NUMERIC].writemask)
|
|
*writemask = hlsl_combine_writemasks(var->regs[HLSL_REGSET_NUMERIC].writemask, *writemask);
|
|
}
|
|
|
|
static void sm4_register_from_deref(struct hlsl_ctx *ctx, struct vkd3d_shader_register *reg,
|
|
uint32_t *writemask, const struct hlsl_deref *deref, struct sm4_instruction *sm4_instr)
|
|
{
|
|
const struct hlsl_type *data_type = hlsl_deref_get_type(ctx, deref);
|
|
const struct hlsl_ir_var *var = deref->var;
|
|
|
|
if (var->is_uniform)
|
|
{
|
|
enum hlsl_regset regset = hlsl_deref_get_regset(ctx, deref);
|
|
|
|
if (regset == HLSL_REGSET_TEXTURES)
|
|
{
|
|
reg->type = VKD3DSPR_RESOURCE;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
if (hlsl_version_ge(ctx, 5, 1))
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_TEXTURES].id;
|
|
reg->idx[1].offset = var->regs[HLSL_REGSET_TEXTURES].index; /* FIXME: array index */
|
|
reg->idx_count = 2;
|
|
}
|
|
else
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_TEXTURES].index;
|
|
reg->idx[0].offset += hlsl_offset_from_deref_safe(ctx, deref);
|
|
reg->idx_count = 1;
|
|
}
|
|
VKD3D_ASSERT(regset == HLSL_REGSET_TEXTURES);
|
|
*writemask = VKD3DSP_WRITEMASK_ALL;
|
|
}
|
|
else if (regset == HLSL_REGSET_UAVS)
|
|
{
|
|
reg->type = VKD3DSPR_UAV;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
if (hlsl_version_ge(ctx, 5, 1))
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_UAVS].id;
|
|
reg->idx[1].offset = var->regs[HLSL_REGSET_UAVS].index; /* FIXME: array index */
|
|
reg->idx_count = 2;
|
|
}
|
|
else
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_UAVS].index;
|
|
reg->idx[0].offset += hlsl_offset_from_deref_safe(ctx, deref);
|
|
reg->idx_count = 1;
|
|
}
|
|
VKD3D_ASSERT(regset == HLSL_REGSET_UAVS);
|
|
*writemask = VKD3DSP_WRITEMASK_ALL;
|
|
}
|
|
else if (regset == HLSL_REGSET_SAMPLERS)
|
|
{
|
|
reg->type = VKD3DSPR_SAMPLER;
|
|
reg->dimension = VSIR_DIMENSION_NONE;
|
|
if (hlsl_version_ge(ctx, 5, 1))
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_SAMPLERS].id;
|
|
reg->idx[1].offset = var->regs[HLSL_REGSET_SAMPLERS].index; /* FIXME: array index */
|
|
reg->idx_count = 2;
|
|
}
|
|
else
|
|
{
|
|
reg->idx[0].offset = var->regs[HLSL_REGSET_SAMPLERS].index;
|
|
reg->idx[0].offset += hlsl_offset_from_deref_safe(ctx, deref);
|
|
reg->idx_count = 1;
|
|
}
|
|
VKD3D_ASSERT(regset == HLSL_REGSET_SAMPLERS);
|
|
*writemask = VKD3DSP_WRITEMASK_ALL;
|
|
}
|
|
else
|
|
{
|
|
unsigned int offset = hlsl_offset_from_deref_safe(ctx, deref) + var->buffer_offset;
|
|
|
|
VKD3D_ASSERT(data_type->class <= HLSL_CLASS_VECTOR);
|
|
reg->type = VKD3DSPR_CONSTBUFFER;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
if (hlsl_version_ge(ctx, 5, 1))
|
|
{
|
|
reg->idx[0].offset = var->buffer->reg.id;
|
|
reg->idx[1].offset = var->buffer->reg.index; /* FIXME: array index */
|
|
reg->idx[2].offset = offset / 4;
|
|
reg->idx_count = 3;
|
|
}
|
|
else
|
|
{
|
|
reg->idx[0].offset = var->buffer->reg.index;
|
|
reg->idx[1].offset = offset / 4;
|
|
reg->idx_count = 2;
|
|
}
|
|
*writemask = ((1u << data_type->dimx) - 1) << (offset & 3);
|
|
}
|
|
}
|
|
else if (var->is_input_semantic)
|
|
{
|
|
bool has_idx;
|
|
|
|
if (hlsl_sm4_register_from_semantic(ctx, &var->semantic, false, ®->type, &has_idx))
|
|
{
|
|
unsigned int offset = hlsl_offset_from_deref_safe(ctx, deref);
|
|
|
|
if (has_idx)
|
|
{
|
|
reg->idx[0].offset = var->semantic.index + offset / 4;
|
|
reg->idx_count = 1;
|
|
}
|
|
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
*writemask = ((1u << data_type->dimx) - 1) << (offset % 4);
|
|
}
|
|
else
|
|
{
|
|
struct hlsl_reg hlsl_reg = hlsl_reg_from_deref(ctx, deref);
|
|
|
|
VKD3D_ASSERT(hlsl_reg.allocated);
|
|
reg->type = VKD3DSPR_INPUT;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
reg->idx[0].offset = hlsl_reg.id;
|
|
reg->idx_count = 1;
|
|
*writemask = hlsl_reg.writemask;
|
|
}
|
|
}
|
|
else if (var->is_output_semantic)
|
|
{
|
|
bool has_idx;
|
|
|
|
if (hlsl_sm4_register_from_semantic(ctx, &var->semantic, true, ®->type, &has_idx))
|
|
{
|
|
unsigned int offset = hlsl_offset_from_deref_safe(ctx, deref);
|
|
|
|
if (has_idx)
|
|
{
|
|
reg->idx[0].offset = var->semantic.index + offset / 4;
|
|
reg->idx_count = 1;
|
|
}
|
|
|
|
if (reg->type == VKD3DSPR_DEPTHOUT)
|
|
reg->dimension = VSIR_DIMENSION_SCALAR;
|
|
else
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
*writemask = ((1u << data_type->dimx) - 1) << (offset % 4);
|
|
}
|
|
else
|
|
{
|
|
struct hlsl_reg hlsl_reg = hlsl_reg_from_deref(ctx, deref);
|
|
|
|
VKD3D_ASSERT(hlsl_reg.allocated);
|
|
reg->type = VKD3DSPR_OUTPUT;
|
|
reg->dimension = VSIR_DIMENSION_VEC4;
|
|
reg->idx[0].offset = hlsl_reg.id;
|
|
reg->idx_count = 1;
|
|
*writemask = hlsl_reg.writemask;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
enum vkd3d_shader_register_type type = deref->var->indexable ? VKD3DSPR_IDXTEMP : VKD3DSPR_TEMP;
|
|
|
|
sm4_numeric_register_from_deref(ctx, reg, type, writemask, deref, sm4_instr);
|
|
}
|
|
}
|
|
|
|
static void sm4_src_from_deref(const struct tpf_compiler *tpf, struct vkd3d_shader_src_param *src,
|
|
const struct hlsl_deref *deref, unsigned int map_writemask, struct sm4_instruction *sm4_instr)
|
|
{
|
|
unsigned int hlsl_swizzle;
|
|
uint32_t writemask;
|
|
|
|
sm4_register_from_deref(tpf->ctx, &src->reg, &writemask, deref, sm4_instr);
|
|
if (vkd3d_sm4_get_default_swizzle_type(&tpf->lookup, src->reg.type) == VKD3D_SM4_SWIZZLE_VEC4)
|
|
{
|
|
hlsl_swizzle = hlsl_map_swizzle(hlsl_swizzle_from_writemask(writemask), map_writemask);
|
|
src->swizzle = swizzle_from_sm4(hlsl_swizzle);
|
|
}
|
|
}
|
|
|
|
static void sm4_dst_from_node(struct vkd3d_shader_dst_param *dst, const struct hlsl_ir_node *instr)
|
|
{
|
|
sm4_register_from_node(&dst->reg, &dst->write_mask, instr);
|
|
}
|
|
|
|
static void sm4_src_from_constant_value(struct vkd3d_shader_src_param *src,
|
|
const struct hlsl_constant_value *value, unsigned int width, unsigned int map_writemask)
|
|
{
|
|
src->swizzle = 0;
|
|
src->reg.type = VKD3DSPR_IMMCONST;
|
|
if (width == 1)
|
|
{
|
|
src->reg.dimension = VSIR_DIMENSION_SCALAR;
|
|
src->reg.u.immconst_u32[0] = value->u[0].u;
|
|
}
|
|
else
|
|
{
|
|
unsigned int i, j = 0;
|
|
|
|
src->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
for (i = 0; i < 4; ++i)
|
|
{
|
|
if ((map_writemask & (1u << i)) && (j < width))
|
|
src->reg.u.immconst_u32[i] = value->u[j++].u;
|
|
else
|
|
src->reg.u.immconst_u32[i] = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void sm4_src_from_node(const struct tpf_compiler *tpf, struct vkd3d_shader_src_param *src,
|
|
const struct hlsl_ir_node *instr, uint32_t map_writemask)
|
|
{
|
|
unsigned int hlsl_swizzle;
|
|
uint32_t writemask;
|
|
|
|
if (instr->type == HLSL_IR_CONSTANT)
|
|
{
|
|
struct hlsl_ir_constant *constant = hlsl_ir_constant(instr);
|
|
|
|
sm4_src_from_constant_value(src, &constant->value, instr->data_type->dimx, map_writemask);
|
|
return;
|
|
}
|
|
|
|
sm4_register_from_node(&src->reg, &writemask, instr);
|
|
if (vkd3d_sm4_get_default_swizzle_type(&tpf->lookup, src->reg.type) == VKD3D_SM4_SWIZZLE_VEC4)
|
|
{
|
|
hlsl_swizzle = hlsl_map_swizzle(hlsl_swizzle_from_writemask(writemask), map_writemask);
|
|
src->swizzle = swizzle_from_sm4(hlsl_swizzle);
|
|
}
|
|
}
|
|
|
|
static unsigned int sm4_get_index_addressing_from_reg(const struct vkd3d_shader_register *reg,
|
|
unsigned int i)
|
|
{
|
|
if (reg->idx[i].rel_addr)
|
|
{
|
|
if (reg->idx[i].offset == 0)
|
|
return VKD3D_SM4_ADDRESSING_RELATIVE;
|
|
else
|
|
return VKD3D_SM4_ADDRESSING_RELATIVE | VKD3D_SM4_ADDRESSING_OFFSET;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static uint32_t sm4_encode_register(const struct tpf_compiler *tpf, const struct vkd3d_shader_register *reg,
|
|
enum vkd3d_sm4_swizzle_type sm4_swizzle_type, uint32_t sm4_swizzle)
|
|
{
|
|
const struct vkd3d_sm4_register_type_info *register_type_info;
|
|
uint32_t sm4_reg_type, sm4_reg_dim;
|
|
uint32_t token = 0;
|
|
|
|
register_type_info = get_info_from_vkd3d_register_type(&tpf->lookup, reg->type);
|
|
if (!register_type_info)
|
|
{
|
|
FIXME("Unhandled vkd3d-shader register type %#x.\n", reg->type);
|
|
sm4_reg_type = VKD3D_SM4_RT_TEMP;
|
|
if (sm4_swizzle_type == VKD3D_SM4_SWIZZLE_DEFAULT)
|
|
sm4_swizzle_type = VKD3D_SM4_SWIZZLE_VEC4;
|
|
}
|
|
else
|
|
{
|
|
sm4_reg_type = register_type_info->sm4_type;
|
|
if (sm4_swizzle_type == VKD3D_SM4_SWIZZLE_DEFAULT)
|
|
sm4_swizzle_type = register_type_info->default_src_swizzle_type;
|
|
}
|
|
sm4_reg_dim = sm4_dimension_from_vsir_dimension(reg->dimension);
|
|
|
|
token |= sm4_reg_type << VKD3D_SM4_REGISTER_TYPE_SHIFT;
|
|
token |= reg->idx_count << VKD3D_SM4_REGISTER_ORDER_SHIFT;
|
|
token |= sm4_reg_dim << VKD3D_SM4_DIMENSION_SHIFT;
|
|
if (reg->idx_count > 0)
|
|
token |= sm4_get_index_addressing_from_reg(reg, 0) << VKD3D_SM4_ADDRESSING_SHIFT0;
|
|
if (reg->idx_count > 1)
|
|
token |= sm4_get_index_addressing_from_reg(reg, 1) << VKD3D_SM4_ADDRESSING_SHIFT1;
|
|
if (reg->idx_count > 2)
|
|
token |= sm4_get_index_addressing_from_reg(reg, 2) << VKD3D_SM4_ADDRESSING_SHIFT2;
|
|
|
|
if (sm4_reg_dim == VKD3D_SM4_DIMENSION_VEC4)
|
|
{
|
|
token |= (uint32_t)sm4_swizzle_type << VKD3D_SM4_SWIZZLE_TYPE_SHIFT;
|
|
|
|
switch (sm4_swizzle_type)
|
|
{
|
|
case VKD3D_SM4_SWIZZLE_NONE:
|
|
VKD3D_ASSERT(sm4_swizzle || register_is_constant(reg));
|
|
token |= (sm4_swizzle << VKD3D_SM4_WRITEMASK_SHIFT) & VKD3D_SM4_WRITEMASK_MASK;
|
|
break;
|
|
|
|
case VKD3D_SM4_SWIZZLE_VEC4:
|
|
token |= (sm4_swizzle << VKD3D_SM4_SWIZZLE_SHIFT) & VKD3D_SM4_SWIZZLE_MASK;
|
|
break;
|
|
|
|
case VKD3D_SM4_SWIZZLE_SCALAR:
|
|
token |= (sm4_swizzle << VKD3D_SM4_SCALAR_DIM_SHIFT) & VKD3D_SM4_SCALAR_DIM_MASK;
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
return token;
|
|
}
|
|
|
|
static void sm4_write_register_index(const struct tpf_compiler *tpf, const struct vkd3d_shader_register *reg,
|
|
unsigned int j)
|
|
{
|
|
unsigned int addressing = sm4_get_index_addressing_from_reg(reg, j);
|
|
struct vkd3d_bytecode_buffer *buffer = tpf->buffer;
|
|
unsigned int k;
|
|
|
|
if (addressing & VKD3D_SM4_ADDRESSING_RELATIVE)
|
|
{
|
|
const struct vkd3d_shader_src_param *idx_src = reg->idx[j].rel_addr;
|
|
uint32_t idx_src_token;
|
|
|
|
VKD3D_ASSERT(idx_src);
|
|
VKD3D_ASSERT(!idx_src->modifiers);
|
|
VKD3D_ASSERT(idx_src->reg.type != VKD3DSPR_IMMCONST);
|
|
idx_src_token = sm4_encode_register(tpf, &idx_src->reg, VKD3D_SM4_SWIZZLE_SCALAR, idx_src->swizzle);
|
|
|
|
put_u32(buffer, idx_src_token);
|
|
for (k = 0; k < idx_src->reg.idx_count; ++k)
|
|
{
|
|
put_u32(buffer, idx_src->reg.idx[k].offset);
|
|
VKD3D_ASSERT(!idx_src->reg.idx[k].rel_addr);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
put_u32(tpf->buffer, reg->idx[j].offset);
|
|
}
|
|
}
|
|
|
|
static void sm4_write_dst_register(const struct tpf_compiler *tpf, const struct vkd3d_shader_dst_param *dst)
|
|
{
|
|
struct vkd3d_bytecode_buffer *buffer = tpf->buffer;
|
|
uint32_t token = 0;
|
|
unsigned int j;
|
|
|
|
token = sm4_encode_register(tpf, &dst->reg, VKD3D_SM4_SWIZZLE_NONE, dst->write_mask);
|
|
put_u32(buffer, token);
|
|
|
|
for (j = 0; j < dst->reg.idx_count; ++j)
|
|
sm4_write_register_index(tpf, &dst->reg, j);
|
|
}
|
|
|
|
static void sm4_write_src_register(const struct tpf_compiler *tpf, const struct vkd3d_shader_src_param *src)
|
|
{
|
|
struct vkd3d_bytecode_buffer *buffer = tpf->buffer;
|
|
uint32_t token = 0, mod_token = 0;
|
|
unsigned int j;
|
|
|
|
token = sm4_encode_register(tpf, &src->reg, VKD3D_SM4_SWIZZLE_DEFAULT, swizzle_to_sm4(src->swizzle));
|
|
|
|
switch (src->modifiers)
|
|
{
|
|
case VKD3DSPSM_NONE:
|
|
mod_token = VKD3D_SM4_REGISTER_MODIFIER_NONE;
|
|
break;
|
|
|
|
case VKD3DSPSM_ABS:
|
|
mod_token = (VKD3D_SM4_REGISTER_MODIFIER_ABS << VKD3D_SM4_REGISTER_MODIFIER_SHIFT)
|
|
| VKD3D_SM4_EXTENDED_OPERAND_MODIFIER;
|
|
break;
|
|
|
|
case VKD3DSPSM_NEG:
|
|
mod_token = (VKD3D_SM4_REGISTER_MODIFIER_NEGATE << VKD3D_SM4_REGISTER_MODIFIER_SHIFT)
|
|
| VKD3D_SM4_EXTENDED_OPERAND_MODIFIER;
|
|
break;
|
|
|
|
case VKD3DSPSM_ABSNEG:
|
|
mod_token = (VKD3D_SM4_REGISTER_MODIFIER_ABS_NEGATE << VKD3D_SM4_REGISTER_MODIFIER_SHIFT)
|
|
| VKD3D_SM4_EXTENDED_OPERAND_MODIFIER;
|
|
break;
|
|
|
|
default:
|
|
ERR("Unhandled register modifier %#x.\n", src->modifiers);
|
|
vkd3d_unreachable();
|
|
break;
|
|
}
|
|
|
|
if (src->modifiers)
|
|
{
|
|
token |= VKD3D_SM4_EXTENDED_OPERAND;
|
|
put_u32(buffer, token);
|
|
put_u32(buffer, mod_token);
|
|
}
|
|
else
|
|
{
|
|
put_u32(buffer, token);
|
|
}
|
|
|
|
for (j = 0; j < src->reg.idx_count; ++j)
|
|
sm4_write_register_index(tpf, &src->reg, j);
|
|
|
|
if (src->reg.type == VKD3DSPR_IMMCONST)
|
|
{
|
|
put_u32(buffer, src->reg.u.immconst_u32[0]);
|
|
if (src->reg.dimension == VSIR_DIMENSION_VEC4)
|
|
{
|
|
put_u32(buffer, src->reg.u.immconst_u32[1]);
|
|
put_u32(buffer, src->reg.u.immconst_u32[2]);
|
|
put_u32(buffer, src->reg.u.immconst_u32[3]);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void sm4_update_stat_counters(const struct tpf_compiler *tpf, const struct sm4_instruction *instr)
|
|
{
|
|
enum vkd3d_shader_type shader_type = tpf->ctx->profile->type;
|
|
enum vkd3d_sm4_stat_field stat_field;
|
|
uint32_t opcode;
|
|
|
|
++tpf->stat->fields[VKD3D_STAT_INSTR_COUNT];
|
|
|
|
opcode = instr->opcode & VKD3D_SM4_OPCODE_MASK;
|
|
stat_field = get_stat_field_from_sm4_opcode(&tpf->lookup, opcode);
|
|
|
|
switch (opcode)
|
|
{
|
|
case VKD3D_SM4_OP_DCL_TEMPS:
|
|
tpf->stat->fields[stat_field] = max(tpf->stat->fields[stat_field], instr->idx[0]);
|
|
break;
|
|
case VKD3D_SM4_OP_DCL_OUTPUT_TOPOLOGY:
|
|
case VKD3D_SM4_OP_DCL_INPUT_PRIMITIVE:
|
|
tpf->stat->fields[stat_field] = (instr->opcode & VKD3D_SM4_PRIMITIVE_TYPE_MASK)
|
|
>> VKD3D_SM4_PRIMITIVE_TYPE_SHIFT;
|
|
break;
|
|
case VKD3D_SM4_OP_DCL_VERTICES_OUT:
|
|
case VKD3D_SM5_OP_DCL_GS_INSTANCES:
|
|
tpf->stat->fields[stat_field] = instr->idx[0];
|
|
break;
|
|
case VKD3D_SM5_OP_DCL_TESSELLATOR_DOMAIN:
|
|
case VKD3D_SM5_OP_DCL_TESSELLATOR_PARTITIONING:
|
|
case VKD3D_SM5_OP_DCL_TESSELLATOR_OUTPUT_PRIMITIVE:
|
|
tpf->stat->fields[stat_field] = (instr->opcode & VKD3D_SM5_TESSELLATOR_MASK) >> VKD3D_SM5_TESSELLATOR_SHIFT;
|
|
break;
|
|
case VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT:
|
|
case VKD3D_SM5_OP_DCL_OUTPUT_CONTROL_POINT_COUNT:
|
|
if ((shader_type == VKD3D_SHADER_TYPE_HULL && opcode == VKD3D_SM5_OP_DCL_OUTPUT_CONTROL_POINT_COUNT)
|
|
|| (shader_type == VKD3D_SHADER_TYPE_DOMAIN
|
|
&& opcode == VKD3D_SM5_OP_DCL_INPUT_CONTROL_POINT_COUNT))
|
|
{
|
|
tpf->stat->fields[stat_field] = (instr->opcode & VKD3D_SM5_CONTROL_POINT_COUNT_MASK)
|
|
>> VKD3D_SM5_CONTROL_POINT_COUNT_SHIFT;
|
|
}
|
|
break;
|
|
default:
|
|
++tpf->stat->fields[stat_field];
|
|
}
|
|
}
|
|
|
|
static void write_sm4_instruction(const struct tpf_compiler *tpf, const struct sm4_instruction *instr)
|
|
{
|
|
uint32_t token = instr->opcode | instr->extra_bits;
|
|
struct vkd3d_bytecode_buffer *buffer = tpf->buffer;
|
|
unsigned int size, i, j;
|
|
size_t token_position;
|
|
|
|
if (instr->modifier_count > 0)
|
|
token |= VKD3D_SM4_INSTRUCTION_MODIFIER;
|
|
|
|
token_position = put_u32(buffer, 0);
|
|
|
|
for (i = 0; i < instr->modifier_count; ++i)
|
|
{
|
|
uint32_t modifier_token = sm4_encode_instruction_modifier(&instr->modifiers[i]);
|
|
|
|
if (instr->modifier_count > i + 1)
|
|
modifier_token |= VKD3D_SM4_INSTRUCTION_MODIFIER;
|
|
put_u32(buffer, modifier_token);
|
|
}
|
|
|
|
for (i = 0; i < instr->dst_count; ++i)
|
|
sm4_write_dst_register(tpf, &instr->dsts[i]);
|
|
|
|
for (i = 0; i < instr->src_count; ++i)
|
|
sm4_write_src_register(tpf, &instr->srcs[i]);
|
|
|
|
if (instr->byte_stride)
|
|
put_u32(buffer, instr->byte_stride);
|
|
|
|
for (j = 0; j < instr->idx_count; ++j)
|
|
put_u32(buffer, instr->idx[j]);
|
|
|
|
size = (bytecode_get_size(buffer) - token_position) / sizeof(uint32_t);
|
|
token |= (size << VKD3D_SM4_INSTRUCTION_LENGTH_SHIFT);
|
|
set_u32(buffer, token_position, token);
|
|
|
|
sm4_update_stat_counters(tpf, instr);
|
|
}
|
|
|
|
static bool encode_texel_offset_as_aoffimmi(struct sm4_instruction *instr,
|
|
const struct hlsl_ir_node *texel_offset)
|
|
{
|
|
struct sm4_instruction_modifier modif;
|
|
struct hlsl_ir_constant *offset;
|
|
|
|
if (!texel_offset || texel_offset->type != HLSL_IR_CONSTANT)
|
|
return false;
|
|
offset = hlsl_ir_constant(texel_offset);
|
|
|
|
modif.type = VKD3D_SM4_MODIFIER_AOFFIMMI;
|
|
modif.u.aoffimmi.u = offset->value.u[0].i;
|
|
modif.u.aoffimmi.v = 0;
|
|
modif.u.aoffimmi.w = 0;
|
|
if (offset->node.data_type->dimx > 1)
|
|
modif.u.aoffimmi.v = offset->value.u[1].i;
|
|
if (offset->node.data_type->dimx > 2)
|
|
modif.u.aoffimmi.w = offset->value.u[2].i;
|
|
if (modif.u.aoffimmi.u < -8 || modif.u.aoffimmi.u > 7
|
|
|| modif.u.aoffimmi.v < -8 || modif.u.aoffimmi.v > 7
|
|
|| modif.u.aoffimmi.w < -8 || modif.u.aoffimmi.w > 7)
|
|
return false;
|
|
|
|
instr->modifiers[instr->modifier_count++] = modif;
|
|
return true;
|
|
}
|
|
|
|
static void write_sm4_dcl_constant_buffer(const struct tpf_compiler *tpf, const struct hlsl_buffer *cbuffer)
|
|
{
|
|
size_t size = (cbuffer->used_size + 3) / 4;
|
|
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_DCL_CONSTANT_BUFFER,
|
|
|
|
.srcs[0].reg.dimension = VSIR_DIMENSION_VEC4,
|
|
.srcs[0].reg.type = VKD3DSPR_CONSTBUFFER,
|
|
.srcs[0].swizzle = VKD3D_SHADER_NO_SWIZZLE,
|
|
.src_count = 1,
|
|
};
|
|
|
|
if (hlsl_version_ge(tpf->ctx, 5, 1))
|
|
{
|
|
instr.srcs[0].reg.idx[0].offset = cbuffer->reg.id;
|
|
instr.srcs[0].reg.idx[1].offset = cbuffer->reg.index;
|
|
instr.srcs[0].reg.idx[2].offset = cbuffer->reg.index; /* FIXME: array end */
|
|
instr.srcs[0].reg.idx_count = 3;
|
|
|
|
instr.idx[0] = size;
|
|
instr.idx[1] = cbuffer->reg.space;
|
|
instr.idx_count = 2;
|
|
}
|
|
else
|
|
{
|
|
instr.srcs[0].reg.idx[0].offset = cbuffer->reg.index;
|
|
instr.srcs[0].reg.idx[1].offset = size;
|
|
instr.srcs[0].reg.idx_count = 2;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_dcl_samplers(const struct tpf_compiler *tpf, const struct extern_resource *resource)
|
|
{
|
|
unsigned int i;
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_DCL_SAMPLER,
|
|
|
|
.dsts[0].reg.type = VKD3DSPR_SAMPLER,
|
|
.dst_count = 1,
|
|
};
|
|
|
|
VKD3D_ASSERT(resource->regset == HLSL_REGSET_SAMPLERS);
|
|
|
|
if (resource->component_type->sampler_dim == HLSL_SAMPLER_DIM_COMPARISON)
|
|
instr.extra_bits |= VKD3D_SM4_SAMPLER_COMPARISON << VKD3D_SM4_SAMPLER_MODE_SHIFT;
|
|
|
|
for (i = 0; i < resource->bind_count; ++i)
|
|
{
|
|
if (resource->var && !resource->var->objects_usage[HLSL_REGSET_SAMPLERS][i].used)
|
|
continue;
|
|
|
|
if (hlsl_version_ge(tpf->ctx, 5, 1))
|
|
{
|
|
VKD3D_ASSERT(!i);
|
|
instr.dsts[0].reg.idx[0].offset = resource->id;
|
|
instr.dsts[0].reg.idx[1].offset = resource->index;
|
|
instr.dsts[0].reg.idx[2].offset = resource->index; /* FIXME: array end */
|
|
instr.dsts[0].reg.idx_count = 3;
|
|
|
|
instr.idx[0] = resource->space;
|
|
instr.idx_count = 1;
|
|
}
|
|
else
|
|
{
|
|
instr.dsts[0].reg.idx[0].offset = resource->index + i;
|
|
instr.dsts[0].reg.idx_count = 1;
|
|
}
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
}
|
|
|
|
static void write_sm4_dcl_textures(const struct tpf_compiler *tpf, const struct extern_resource *resource,
|
|
bool uav)
|
|
{
|
|
enum hlsl_regset regset = uav ? HLSL_REGSET_UAVS : HLSL_REGSET_TEXTURES;
|
|
struct hlsl_type *component_type;
|
|
struct sm4_instruction instr;
|
|
bool multisampled;
|
|
unsigned int i;
|
|
|
|
VKD3D_ASSERT(resource->regset == regset);
|
|
|
|
component_type = resource->component_type;
|
|
|
|
for (i = 0; i < resource->bind_count; ++i)
|
|
{
|
|
if (resource->var && !resource->var->objects_usage[regset][i].used)
|
|
continue;
|
|
|
|
instr = (struct sm4_instruction)
|
|
{
|
|
.dsts[0].reg.type = uav ? VKD3DSPR_UAV : VKD3DSPR_RESOURCE,
|
|
.dsts[0].reg.idx[0].offset = resource->id + i,
|
|
.dsts[0].reg.idx_count = 1,
|
|
.dst_count = 1,
|
|
|
|
.idx[0] = sm4_data_type(component_type) * 0x1111,
|
|
.idx_count = 1,
|
|
};
|
|
|
|
multisampled = component_type->sampler_dim == HLSL_SAMPLER_DIM_2DMS
|
|
|| component_type->sampler_dim == HLSL_SAMPLER_DIM_2DMSARRAY;
|
|
|
|
if (hlsl_version_lt(tpf->ctx, 4, 1) && multisampled && !component_type->sample_count)
|
|
{
|
|
hlsl_error(tpf->ctx, &resource->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TYPE,
|
|
"Multisampled texture object declaration needs sample count for profile %s.",
|
|
tpf->ctx->profile->name);
|
|
}
|
|
|
|
if (hlsl_version_ge(tpf->ctx, 5, 1))
|
|
{
|
|
VKD3D_ASSERT(!i);
|
|
instr.dsts[0].reg.idx[0].offset = resource->id;
|
|
instr.dsts[0].reg.idx[1].offset = resource->index;
|
|
instr.dsts[0].reg.idx[2].offset = resource->index; /* FIXME: array end */
|
|
instr.dsts[0].reg.idx_count = 3;
|
|
|
|
instr.idx[1] = resource->space;
|
|
instr.idx_count = 2;
|
|
}
|
|
else
|
|
{
|
|
instr.dsts[0].reg.idx[0].offset = resource->index + i;
|
|
instr.dsts[0].reg.idx_count = 1;
|
|
}
|
|
|
|
if (uav)
|
|
{
|
|
switch (component_type->sampler_dim)
|
|
{
|
|
case HLSL_SAMPLER_DIM_STRUCTURED_BUFFER:
|
|
instr.opcode = VKD3D_SM5_OP_DCL_UAV_STRUCTURED;
|
|
instr.byte_stride = component_type->e.resource.format->reg_size[HLSL_REGSET_NUMERIC] * 4;
|
|
break;
|
|
default:
|
|
instr.opcode = VKD3D_SM5_OP_DCL_UAV_TYPED;
|
|
break;
|
|
}
|
|
|
|
if (component_type->e.resource.rasteriser_ordered)
|
|
instr.opcode |= VKD3DSUF_RASTERISER_ORDERED_VIEW << VKD3D_SM5_UAV_FLAGS_SHIFT;
|
|
}
|
|
else
|
|
{
|
|
instr.opcode = VKD3D_SM4_OP_DCL_RESOURCE;
|
|
}
|
|
instr.extra_bits |= (sm4_resource_dimension(component_type) << VKD3D_SM4_RESOURCE_TYPE_SHIFT);
|
|
|
|
if (multisampled)
|
|
instr.extra_bits |= component_type->sample_count << VKD3D_SM4_RESOURCE_SAMPLE_COUNT_SHIFT;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
}
|
|
|
|
static void write_sm4_dcl_semantic(const struct tpf_compiler *tpf, const struct hlsl_ir_var *var)
|
|
{
|
|
const struct hlsl_profile_info *profile = tpf->ctx->profile;
|
|
const bool output = var->is_output_semantic;
|
|
enum vkd3d_shader_sysval_semantic semantic;
|
|
bool has_idx;
|
|
|
|
struct sm4_instruction instr =
|
|
{
|
|
.dsts[0].reg.dimension = VSIR_DIMENSION_VEC4,
|
|
.dst_count = 1,
|
|
};
|
|
|
|
if (hlsl_sm4_register_from_semantic(tpf->ctx, &var->semantic, output, &instr.dsts[0].reg.type, &has_idx))
|
|
{
|
|
if (has_idx)
|
|
{
|
|
instr.dsts[0].reg.idx[0].offset = var->semantic.index;
|
|
instr.dsts[0].reg.idx_count = 1;
|
|
}
|
|
else
|
|
{
|
|
instr.dsts[0].reg.idx_count = 0;
|
|
}
|
|
instr.dsts[0].write_mask = (1 << var->data_type->dimx) - 1;
|
|
}
|
|
else
|
|
{
|
|
instr.dsts[0].reg.type = output ? VKD3DSPR_OUTPUT : VKD3DSPR_INPUT;
|
|
instr.dsts[0].reg.idx[0].offset = var->regs[HLSL_REGSET_NUMERIC].id;
|
|
instr.dsts[0].reg.idx_count = 1;
|
|
instr.dsts[0].write_mask = var->regs[HLSL_REGSET_NUMERIC].writemask;
|
|
}
|
|
|
|
if (instr.dsts[0].reg.type == VKD3DSPR_DEPTHOUT)
|
|
instr.dsts[0].reg.dimension = VSIR_DIMENSION_SCALAR;
|
|
|
|
sysval_semantic_from_hlsl(&semantic, tpf->ctx, &var->semantic, output);
|
|
if (semantic == ~0u)
|
|
semantic = VKD3D_SHADER_SV_NONE;
|
|
|
|
if (var->is_input_semantic)
|
|
{
|
|
switch (semantic)
|
|
{
|
|
case VKD3D_SHADER_SV_NONE:
|
|
instr.opcode = (profile->type == VKD3D_SHADER_TYPE_PIXEL)
|
|
? VKD3D_SM4_OP_DCL_INPUT_PS : VKD3D_SM4_OP_DCL_INPUT;
|
|
break;
|
|
|
|
case VKD3D_SHADER_SV_INSTANCE_ID:
|
|
case VKD3D_SHADER_SV_IS_FRONT_FACE:
|
|
case VKD3D_SHADER_SV_PRIMITIVE_ID:
|
|
case VKD3D_SHADER_SV_SAMPLE_INDEX:
|
|
case VKD3D_SHADER_SV_VERTEX_ID:
|
|
instr.opcode = (profile->type == VKD3D_SHADER_TYPE_PIXEL)
|
|
? VKD3D_SM4_OP_DCL_INPUT_PS_SGV : VKD3D_SM4_OP_DCL_INPUT_SGV;
|
|
break;
|
|
|
|
default:
|
|
instr.opcode = (profile->type == VKD3D_SHADER_TYPE_PIXEL)
|
|
? VKD3D_SM4_OP_DCL_INPUT_PS_SIV : VKD3D_SM4_OP_DCL_INPUT_SIV;
|
|
break;
|
|
}
|
|
|
|
if (profile->type == VKD3D_SHADER_TYPE_PIXEL)
|
|
{
|
|
enum vkd3d_shader_interpolation_mode mode = VKD3DSIM_LINEAR;
|
|
|
|
if ((var->storage_modifiers & HLSL_STORAGE_NOINTERPOLATION) || type_is_integer(var->data_type))
|
|
{
|
|
mode = VKD3DSIM_CONSTANT;
|
|
}
|
|
else
|
|
{
|
|
static const struct
|
|
{
|
|
unsigned int modifiers;
|
|
enum vkd3d_shader_interpolation_mode mode;
|
|
}
|
|
modes[] =
|
|
{
|
|
{ HLSL_STORAGE_CENTROID | HLSL_STORAGE_NOPERSPECTIVE, VKD3DSIM_LINEAR_NOPERSPECTIVE_CENTROID },
|
|
{ HLSL_STORAGE_NOPERSPECTIVE, VKD3DSIM_LINEAR_NOPERSPECTIVE },
|
|
{ HLSL_STORAGE_CENTROID, VKD3DSIM_LINEAR_CENTROID },
|
|
{ HLSL_STORAGE_CENTROID | HLSL_STORAGE_LINEAR, VKD3DSIM_LINEAR_CENTROID },
|
|
};
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(modes); ++i)
|
|
{
|
|
if ((var->storage_modifiers & modes[i].modifiers) == modes[i].modifiers)
|
|
{
|
|
mode = modes[i].mode;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
instr.extra_bits |= mode << VKD3D_SM4_INTERPOLATION_MODE_SHIFT;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (semantic == VKD3D_SHADER_SV_NONE || profile->type == VKD3D_SHADER_TYPE_PIXEL)
|
|
instr.opcode = VKD3D_SM4_OP_DCL_OUTPUT;
|
|
else
|
|
instr.opcode = VKD3D_SM4_OP_DCL_OUTPUT_SIV;
|
|
}
|
|
|
|
if (instr.opcode == VKD3D_SM4_OP_DCL_OUTPUT)
|
|
{
|
|
VKD3D_ASSERT(semantic == VKD3D_SHADER_SV_NONE || semantic == VKD3D_SHADER_SV_TARGET
|
|
|| instr.dsts[0].reg.type != VKD3DSPR_OUTPUT);
|
|
}
|
|
else if (instr.opcode == VKD3D_SM4_OP_DCL_INPUT || instr.opcode == VKD3D_SM4_OP_DCL_INPUT_PS)
|
|
{
|
|
VKD3D_ASSERT(semantic == VKD3D_SHADER_SV_NONE);
|
|
}
|
|
else
|
|
{
|
|
VKD3D_ASSERT(semantic != VKD3D_SHADER_SV_NONE);
|
|
instr.idx_count = 1;
|
|
instr.idx[0] = semantic;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_dcl_temps(const struct tpf_compiler *tpf, uint32_t temp_count)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_DCL_TEMPS,
|
|
|
|
.idx = {temp_count},
|
|
.idx_count = 1,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_dcl_indexable_temp(const struct tpf_compiler *tpf, uint32_t idx,
|
|
uint32_t size, uint32_t comp_count)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_DCL_INDEXABLE_TEMP,
|
|
|
|
.idx = {idx, size, comp_count},
|
|
.idx_count = 3,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_dcl_thread_group(const struct tpf_compiler *tpf, const uint32_t thread_count[3])
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM5_OP_DCL_THREAD_GROUP,
|
|
|
|
.idx[0] = thread_count[0],
|
|
.idx[1] = thread_count[1],
|
|
.idx[2] = thread_count[2],
|
|
.idx_count = 3,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_dcl_global_flags(const struct tpf_compiler *tpf, uint32_t flags)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_DCL_GLOBAL_FLAGS,
|
|
.extra_bits = flags << VKD3D_SM4_GLOBAL_FLAGS_SHIFT,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_ret(const struct tpf_compiler *tpf)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_RET,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_unary_op(const struct tpf_compiler *tpf, enum vkd3d_sm4_opcode opcode,
|
|
const struct hlsl_ir_node *dst, const struct hlsl_ir_node *src, enum vkd3d_shader_src_modifier src_mod)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src, instr.dsts[0].write_mask);
|
|
instr.srcs[0].modifiers = src_mod;
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_unary_op_with_two_destinations(const struct tpf_compiler *tpf, enum vkd3d_sm4_opcode opcode,
|
|
const struct hlsl_ir_node *dst, unsigned int dst_idx, const struct hlsl_ir_node *src)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
VKD3D_ASSERT(dst_idx < ARRAY_SIZE(instr.dsts));
|
|
sm4_dst_from_node(&instr.dsts[dst_idx], dst);
|
|
instr.dsts[1 - dst_idx].reg.type = VKD3DSPR_NULL;
|
|
instr.dsts[1 - dst_idx].reg.dimension = VSIR_DIMENSION_NONE;
|
|
instr.dsts[1 - dst_idx].reg.idx_count = 0;
|
|
instr.dst_count = 2;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src, instr.dsts[dst_idx].write_mask);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_binary_op(const struct tpf_compiler *tpf, enum vkd3d_sm4_opcode opcode,
|
|
const struct hlsl_ir_node *dst, const struct hlsl_ir_node *src1, const struct hlsl_ir_node *src2)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src1, instr.dsts[0].write_mask);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], src2, instr.dsts[0].write_mask);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
/* dp# instructions don't map the swizzle. */
|
|
static void write_sm4_binary_op_dot(const struct tpf_compiler *tpf, enum vkd3d_sm4_opcode opcode,
|
|
const struct hlsl_ir_node *dst, const struct hlsl_ir_node *src1, const struct hlsl_ir_node *src2)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src1, VKD3DSP_WRITEMASK_ALL);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], src2, VKD3DSP_WRITEMASK_ALL);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_binary_op_with_two_destinations(const struct tpf_compiler *tpf,
|
|
enum vkd3d_sm4_opcode opcode, const struct hlsl_ir_node *dst, unsigned int dst_idx,
|
|
const struct hlsl_ir_node *src1, const struct hlsl_ir_node *src2)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
VKD3D_ASSERT(dst_idx < ARRAY_SIZE(instr.dsts));
|
|
sm4_dst_from_node(&instr.dsts[dst_idx], dst);
|
|
instr.dsts[1 - dst_idx].reg.type = VKD3DSPR_NULL;
|
|
instr.dsts[1 - dst_idx].reg.dimension = VSIR_DIMENSION_NONE;
|
|
instr.dsts[1 - dst_idx].reg.idx_count = 0;
|
|
instr.dst_count = 2;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src1, instr.dsts[dst_idx].write_mask);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], src2, instr.dsts[dst_idx].write_mask);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_ternary_op(const struct tpf_compiler *tpf, enum vkd3d_sm4_opcode opcode,
|
|
const struct hlsl_ir_node *dst, const struct hlsl_ir_node *src1, const struct hlsl_ir_node *src2,
|
|
const struct hlsl_ir_node *src3)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = opcode;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], src1, instr.dsts[0].write_mask);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], src2, instr.dsts[0].write_mask);
|
|
sm4_src_from_node(tpf, &instr.srcs[2], src3, instr.dsts[0].write_mask);
|
|
instr.src_count = 3;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_ld(const struct tpf_compiler *tpf, const struct hlsl_ir_node *dst,
|
|
const struct hlsl_deref *resource, const struct hlsl_ir_node *coords,
|
|
const struct hlsl_ir_node *sample_index, const struct hlsl_ir_node *texel_offset,
|
|
enum hlsl_sampler_dim dim)
|
|
{
|
|
const struct hlsl_type *resource_type = hlsl_deref_get_type(tpf->ctx, resource);
|
|
bool multisampled = resource_type->class == HLSL_CLASS_TEXTURE
|
|
&& (resource_type->sampler_dim == HLSL_SAMPLER_DIM_2DMS || resource_type->sampler_dim == HLSL_SAMPLER_DIM_2DMSARRAY);
|
|
bool uav = (hlsl_deref_get_regset(tpf->ctx, resource) == HLSL_REGSET_UAVS);
|
|
unsigned int coords_writemask = VKD3DSP_WRITEMASK_ALL;
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
if (uav)
|
|
instr.opcode = VKD3D_SM5_OP_LD_UAV_TYPED;
|
|
else
|
|
instr.opcode = multisampled ? VKD3D_SM4_OP_LD2DMS : VKD3D_SM4_OP_LD;
|
|
|
|
if (texel_offset)
|
|
{
|
|
if (!encode_texel_offset_as_aoffimmi(&instr, texel_offset))
|
|
{
|
|
hlsl_error(tpf->ctx, &texel_offset->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TEXEL_OFFSET,
|
|
"Offset must resolve to integer literal in the range -8 to 7.");
|
|
return;
|
|
}
|
|
}
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
if (!uav)
|
|
{
|
|
/* Mipmap level is in the last component in the IR, but needs to be in the W
|
|
* component in the instruction. */
|
|
unsigned int dim_count = hlsl_sampler_dim_count(dim);
|
|
|
|
if (dim_count == 1)
|
|
coords_writemask = VKD3DSP_WRITEMASK_0 | VKD3DSP_WRITEMASK_3;
|
|
if (dim_count == 2)
|
|
coords_writemask = VKD3DSP_WRITEMASK_0 | VKD3DSP_WRITEMASK_1 | VKD3DSP_WRITEMASK_3;
|
|
}
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], coords, coords_writemask);
|
|
|
|
sm4_src_from_deref(tpf, &instr.srcs[1], resource, instr.dsts[0].write_mask, &instr);
|
|
|
|
instr.src_count = 2;
|
|
|
|
if (multisampled)
|
|
{
|
|
if (sample_index->type == HLSL_IR_CONSTANT)
|
|
{
|
|
struct vkd3d_shader_register *reg = &instr.srcs[2].reg;
|
|
struct hlsl_ir_constant *index;
|
|
|
|
index = hlsl_ir_constant(sample_index);
|
|
|
|
memset(&instr.srcs[2], 0, sizeof(instr.srcs[2]));
|
|
reg->type = VKD3DSPR_IMMCONST;
|
|
reg->dimension = VSIR_DIMENSION_SCALAR;
|
|
reg->u.immconst_u32[0] = index->value.u[0].u;
|
|
}
|
|
else if (tpf->ctx->profile->major_version == 4 && tpf->ctx->profile->minor_version == 0)
|
|
{
|
|
hlsl_error(tpf->ctx, &sample_index->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TYPE, "Expected literal sample index.");
|
|
}
|
|
else
|
|
{
|
|
sm4_src_from_node(tpf, &instr.srcs[2], sample_index, 0);
|
|
}
|
|
|
|
++instr.src_count;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_sample(const struct tpf_compiler *tpf, const struct hlsl_ir_resource_load *load)
|
|
{
|
|
const struct hlsl_ir_node *texel_offset = load->texel_offset.node;
|
|
const struct hlsl_ir_node *coords = load->coords.node;
|
|
const struct hlsl_deref *resource = &load->resource;
|
|
const struct hlsl_deref *sampler = &load->sampler;
|
|
const struct hlsl_ir_node *dst = &load->node;
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
switch (load->load_type)
|
|
{
|
|
case HLSL_RESOURCE_SAMPLE:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE;
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_CMP:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_C;
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_CMP_LZ:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_C_LZ;
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_LOD:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_LOD;
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_LOD_BIAS:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_B;
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_GRAD:
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_GRAD;
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
|
|
if (texel_offset)
|
|
{
|
|
if (!encode_texel_offset_as_aoffimmi(&instr, texel_offset))
|
|
{
|
|
hlsl_error(tpf->ctx, &texel_offset->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TEXEL_OFFSET,
|
|
"Offset must resolve to integer literal in the range -8 to 7.");
|
|
return;
|
|
}
|
|
}
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], coords, VKD3DSP_WRITEMASK_ALL);
|
|
sm4_src_from_deref(tpf, &instr.srcs[1], resource, instr.dsts[0].write_mask, &instr);
|
|
sm4_src_from_deref(tpf, &instr.srcs[2], sampler, VKD3DSP_WRITEMASK_ALL, &instr);
|
|
instr.src_count = 3;
|
|
|
|
if (load->load_type == HLSL_RESOURCE_SAMPLE_LOD
|
|
|| load->load_type == HLSL_RESOURCE_SAMPLE_LOD_BIAS)
|
|
{
|
|
sm4_src_from_node(tpf, &instr.srcs[3], load->lod.node, VKD3DSP_WRITEMASK_ALL);
|
|
++instr.src_count;
|
|
}
|
|
else if (load->load_type == HLSL_RESOURCE_SAMPLE_GRAD)
|
|
{
|
|
sm4_src_from_node(tpf, &instr.srcs[3], load->ddx.node, VKD3DSP_WRITEMASK_ALL);
|
|
sm4_src_from_node(tpf, &instr.srcs[4], load->ddy.node, VKD3DSP_WRITEMASK_ALL);
|
|
instr.src_count += 2;
|
|
}
|
|
else if (load->load_type == HLSL_RESOURCE_SAMPLE_CMP
|
|
|| load->load_type == HLSL_RESOURCE_SAMPLE_CMP_LZ)
|
|
{
|
|
sm4_src_from_node(tpf, &instr.srcs[3], load->cmp.node, VKD3DSP_WRITEMASK_ALL);
|
|
++instr.src_count;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_sampleinfo(const struct tpf_compiler *tpf, const struct hlsl_ir_resource_load *load)
|
|
{
|
|
const struct hlsl_deref *resource = &load->resource;
|
|
const struct hlsl_ir_node *dst = &load->node;
|
|
struct sm4_instruction instr;
|
|
|
|
VKD3D_ASSERT(dst->data_type->e.numeric.type == HLSL_TYPE_UINT || dst->data_type->e.numeric.type == HLSL_TYPE_FLOAT);
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_INFO;
|
|
if (dst->data_type->e.numeric.type == HLSL_TYPE_UINT)
|
|
instr.extra_bits |= VKD3DSI_SAMPLE_INFO_UINT << VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_deref(tpf, &instr.srcs[0], resource, instr.dsts[0].write_mask, &instr);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_resinfo(const struct tpf_compiler *tpf, const struct hlsl_ir_resource_load *load)
|
|
{
|
|
const struct hlsl_deref *resource = &load->resource;
|
|
const struct hlsl_ir_node *dst = &load->node;
|
|
struct sm4_instruction instr;
|
|
|
|
if (resource->data_type->sampler_dim == HLSL_SAMPLER_DIM_BUFFER
|
|
|| resource->data_type->sampler_dim == HLSL_SAMPLER_DIM_STRUCTURED_BUFFER)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &load->node.loc, "resinfo for buffers.");
|
|
return;
|
|
}
|
|
|
|
VKD3D_ASSERT(dst->data_type->e.numeric.type == HLSL_TYPE_UINT || dst->data_type->e.numeric.type == HLSL_TYPE_FLOAT);
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_RESINFO;
|
|
if (dst->data_type->e.numeric.type == HLSL_TYPE_UINT)
|
|
instr.extra_bits |= VKD3DSI_RESINFO_UINT << VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], load->lod.node, VKD3DSP_WRITEMASK_ALL);
|
|
sm4_src_from_deref(tpf, &instr.srcs[1], resource, instr.dsts[0].write_mask, &instr);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static bool type_is_float(const struct hlsl_type *type)
|
|
{
|
|
return type->e.numeric.type == HLSL_TYPE_FLOAT || type->e.numeric.type == HLSL_TYPE_HALF;
|
|
}
|
|
|
|
static void write_sm4_cast_from_bool(const struct tpf_compiler *tpf, const struct hlsl_ir_expr *expr,
|
|
const struct hlsl_ir_node *arg, uint32_t mask)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_AND;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], &expr->node);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], arg, instr.dsts[0].write_mask);
|
|
instr.srcs[1].reg.type = VKD3DSPR_IMMCONST;
|
|
instr.srcs[1].reg.dimension = VSIR_DIMENSION_SCALAR;
|
|
instr.srcs[1].reg.u.immconst_u32[0] = mask;
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_cast(const struct tpf_compiler *tpf, const struct hlsl_ir_expr *expr)
|
|
{
|
|
static const union
|
|
{
|
|
uint32_t u;
|
|
float f;
|
|
} one = { .f = 1.0 };
|
|
const struct hlsl_ir_node *arg1 = expr->operands[0].node;
|
|
const struct hlsl_type *dst_type = expr->node.data_type;
|
|
const struct hlsl_type *src_type = arg1->data_type;
|
|
|
|
/* Narrowing casts were already lowered. */
|
|
VKD3D_ASSERT(src_type->dimx == dst_type->dimx);
|
|
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_HALF:
|
|
case HLSL_TYPE_FLOAT:
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_HALF:
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_ITOF, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_UTOF, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
write_sm4_cast_from_bool(tpf, expr, arg1, one.u);
|
|
break;
|
|
|
|
case HLSL_TYPE_DOUBLE:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 cast from double to float.");
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_HALF:
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_FTOI, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
write_sm4_cast_from_bool(tpf, expr, arg1, 1);
|
|
break;
|
|
|
|
case HLSL_TYPE_DOUBLE:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 cast from double to int.");
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
break;
|
|
|
|
case HLSL_TYPE_UINT:
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_HALF:
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_FTOU, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
write_sm4_cast_from_bool(tpf, expr, arg1, 1);
|
|
break;
|
|
|
|
case HLSL_TYPE_DOUBLE:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 cast from double to uint.");
|
|
break;
|
|
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
break;
|
|
|
|
case HLSL_TYPE_DOUBLE:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 cast to double.");
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
/* Casts to bool should have already been lowered. */
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void write_sm4_store_uav_typed(const struct tpf_compiler *tpf, const struct hlsl_deref *dst,
|
|
const struct hlsl_ir_node *coords, const struct hlsl_ir_node *value)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM5_OP_STORE_UAV_TYPED;
|
|
|
|
sm4_register_from_deref(tpf->ctx, &instr.dsts[0].reg, &instr.dsts[0].write_mask, dst, &instr);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], coords, VKD3DSP_WRITEMASK_ALL);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], value, VKD3DSP_WRITEMASK_ALL);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_rasterizer_sample_count(const struct tpf_compiler *tpf, const struct hlsl_ir_node *dst)
|
|
{
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_SAMPLE_INFO;
|
|
instr.extra_bits |= VKD3DSI_SAMPLE_INFO_UINT << VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
instr.srcs[0].reg.type = VKD3DSPR_RASTERIZER;
|
|
instr.srcs[0].reg.dimension = VSIR_DIMENSION_VEC4;
|
|
instr.srcs[0].swizzle = VKD3D_SHADER_SWIZZLE(X, X, X, X);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_expr(const struct tpf_compiler *tpf, const struct hlsl_ir_expr *expr)
|
|
{
|
|
const struct hlsl_ir_node *arg1 = expr->operands[0].node;
|
|
const struct hlsl_ir_node *arg2 = expr->operands[1].node;
|
|
const struct hlsl_ir_node *arg3 = expr->operands[2].node;
|
|
const struct hlsl_type *dst_type = expr->node.data_type;
|
|
struct vkd3d_string_buffer *dst_type_string;
|
|
|
|
VKD3D_ASSERT(expr->node.reg.allocated);
|
|
|
|
if (!(dst_type_string = hlsl_type_to_string(tpf->ctx, dst_type)))
|
|
return;
|
|
|
|
switch (expr->op)
|
|
{
|
|
case HLSL_OP0_RASTERIZER_SAMPLE_COUNT:
|
|
if (tpf->ctx->profile->type == VKD3D_SHADER_TYPE_PIXEL && hlsl_version_ge(tpf->ctx, 4, 1))
|
|
write_sm4_rasterizer_sample_count(tpf, &expr->node);
|
|
else
|
|
hlsl_error(tpf->ctx, &expr->node.loc, VKD3D_SHADER_ERROR_HLSL_INCOMPATIBLE_PROFILE,
|
|
"GetRenderTargetSampleCount() can only be used from a pixel shader using version 4.1 or higher.");
|
|
break;
|
|
|
|
case HLSL_OP1_ABS:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, VKD3DSPSM_ABS);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s absolute value expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP1_BIT_NOT:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_NOT, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_CAST:
|
|
write_sm4_cast(tpf, expr);
|
|
break;
|
|
|
|
case HLSL_OP1_CEIL:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_ROUND_PI, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_COS:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op_with_two_destinations(tpf, VKD3D_SM4_OP_SINCOS, &expr->node, 1, arg1);
|
|
break;
|
|
|
|
case HLSL_OP1_DSX:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_DERIV_RTX, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_DSX_COARSE:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_DERIV_RTX_COARSE, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_DSX_FINE:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_DERIV_RTX_FINE, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_DSY:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_DERIV_RTY, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_DSY_COARSE:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_DERIV_RTY_COARSE, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_DSY_FINE:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_DERIV_RTY_FINE, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_EXP2:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_EXP, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_F16TOF32:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
VKD3D_ASSERT(hlsl_version_ge(tpf->ctx, 5, 0));
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_F16TOF32, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_FLOOR:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_ROUND_NI, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_FRACT:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_FRC, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_LOG2:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_LOG, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_LOGIC_NOT:
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_NOT, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_NEG:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, VKD3DSPSM_NEG);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_INEG, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s negation expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP1_RCP:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
/* SM5 comes with a RCP opcode */
|
|
if (tpf->ctx->profile->major_version >= 5)
|
|
{
|
|
write_sm4_unary_op(tpf, VKD3D_SM5_OP_RCP, &expr->node, arg1, 0);
|
|
}
|
|
else
|
|
{
|
|
/* For SM4, implement as DIV dst, 1.0, src */
|
|
struct sm4_instruction instr;
|
|
struct hlsl_constant_value one;
|
|
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_DIV;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], &expr->node);
|
|
instr.dst_count = 1;
|
|
|
|
for (unsigned int i = 0; i < 4; i++)
|
|
one.u[i].f = 1.0f;
|
|
sm4_src_from_constant_value(&instr.srcs[0], &one, dst_type->dimx, instr.dsts[0].write_mask);
|
|
sm4_src_from_node(tpf, &instr.srcs[1], arg1, instr.dsts[0].write_mask);
|
|
instr.src_count = 2;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s rcp expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP1_REINTERPRET:
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_ROUND:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_ROUND_NE, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_RSQ:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_RSQ, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_SAT:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_MOV
|
|
| (VKD3D_SM4_INSTRUCTION_FLAG_SATURATE << VKD3D_SM4_INSTRUCTION_FLAGS_SHIFT),
|
|
&expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_SIN:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op_with_two_destinations(tpf, VKD3D_SM4_OP_SINCOS, &expr->node, 0, arg1);
|
|
break;
|
|
|
|
case HLSL_OP1_SQRT:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_SQRT, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP1_TRUNC:
|
|
VKD3D_ASSERT(type_is_float(dst_type));
|
|
write_sm4_unary_op(tpf, VKD3D_SM4_OP_ROUND_Z, &expr->node, arg1, 0);
|
|
break;
|
|
|
|
case HLSL_OP2_ADD:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_ADD, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_IADD, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s addition expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_BIT_AND:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_AND, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_BIT_OR:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_OR, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_BIT_XOR:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_XOR, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_DIV:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_DIV, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op_with_two_destinations(tpf, VKD3D_SM4_OP_UDIV, &expr->node, 0, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s division expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_DOT:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
switch (arg1->data_type->dimx)
|
|
{
|
|
case 4:
|
|
write_sm4_binary_op_dot(tpf, VKD3D_SM4_OP_DP4, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case 3:
|
|
write_sm4_binary_op_dot(tpf, VKD3D_SM4_OP_DP3, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case 2:
|
|
write_sm4_binary_op_dot(tpf, VKD3D_SM4_OP_DP2, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case 1:
|
|
default:
|
|
vkd3d_unreachable();
|
|
}
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s dot expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_EQUAL:
|
|
{
|
|
const struct hlsl_type *src_type = arg1->data_type;
|
|
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_EQ, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_IEQ, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 equality between \"%s\" operands.",
|
|
debug_hlsl_type(tpf->ctx, src_type));
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case HLSL_OP2_GEQUAL:
|
|
{
|
|
const struct hlsl_type *src_type = arg1->data_type;
|
|
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_GE, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_IGE, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_UGE, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 greater-than-or-equal between \"%s\" operands.",
|
|
debug_hlsl_type(tpf->ctx, src_type));
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case HLSL_OP2_LESS:
|
|
{
|
|
const struct hlsl_type *src_type = arg1->data_type;
|
|
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_LT, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_ILT, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_ULT, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 less-than between \"%s\" operands.",
|
|
debug_hlsl_type(tpf->ctx, src_type));
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case HLSL_OP2_LOGIC_AND:
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_AND, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_LOGIC_OR:
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_OR, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_LSHIFT:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
VKD3D_ASSERT(dst_type->e.numeric.type != HLSL_TYPE_BOOL);
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_ISHL, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP2_MAX:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_MAX, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_IMAX, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_UMAX, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s maximum expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_MIN:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_MIN, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_IMIN, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_UMIN, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s minimum expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_MOD:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op_with_two_destinations(tpf, VKD3D_SM4_OP_UDIV, &expr->node, 1, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s modulus expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_MUL:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_MUL, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
/* Using IMUL instead of UMUL because we're taking the low
|
|
* bits, and the native compiler generates IMUL. */
|
|
write_sm4_binary_op_with_two_destinations(tpf, VKD3D_SM4_OP_IMUL, &expr->node, 1, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s multiplication expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
case HLSL_OP2_NEQUAL:
|
|
{
|
|
const struct hlsl_type *src_type = arg1->data_type;
|
|
|
|
VKD3D_ASSERT(dst_type->e.numeric.type == HLSL_TYPE_BOOL);
|
|
|
|
switch (src_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_NE, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_TYPE_BOOL:
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_binary_op(tpf, VKD3D_SM4_OP_INE, &expr->node, arg1, arg2);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 inequality between \"%s\" operands.",
|
|
debug_hlsl_type(tpf->ctx, src_type));
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
|
|
case HLSL_OP2_RSHIFT:
|
|
VKD3D_ASSERT(type_is_integer(dst_type));
|
|
VKD3D_ASSERT(dst_type->e.numeric.type != HLSL_TYPE_BOOL);
|
|
write_sm4_binary_op(tpf, dst_type->e.numeric.type == HLSL_TYPE_INT ? VKD3D_SM4_OP_ISHR : VKD3D_SM4_OP_USHR,
|
|
&expr->node, arg1, arg2);
|
|
break;
|
|
|
|
case HLSL_OP3_TERNARY:
|
|
write_sm4_ternary_op(tpf, VKD3D_SM4_OP_MOVC, &expr->node, arg1, arg2, arg3);
|
|
break;
|
|
|
|
case HLSL_OP3_MAD:
|
|
switch (dst_type->e.numeric.type)
|
|
{
|
|
case HLSL_TYPE_FLOAT:
|
|
write_sm4_ternary_op(tpf, VKD3D_SM4_OP_MAD, &expr->node, arg1, arg2, arg3);
|
|
break;
|
|
|
|
case HLSL_TYPE_INT:
|
|
case HLSL_TYPE_UINT:
|
|
write_sm4_ternary_op(tpf, VKD3D_SM4_OP_IMAD, &expr->node, arg1, arg2, arg3);
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s negation expression.", dst_type_string->buffer);
|
|
}
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &expr->node.loc, "SM4 %s expression.", debug_hlsl_expr_op(expr->op));
|
|
}
|
|
|
|
hlsl_release_string_buffer(tpf->ctx, dst_type_string);
|
|
}
|
|
|
|
static void write_sm4_if(const struct tpf_compiler *tpf, const struct hlsl_ir_if *iff)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_IF | VKD3D_SM4_CONDITIONAL_NZ,
|
|
.src_count = 1,
|
|
};
|
|
|
|
VKD3D_ASSERT(iff->condition.node->data_type->dimx == 1);
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], iff->condition.node, VKD3DSP_WRITEMASK_ALL);
|
|
write_sm4_instruction(tpf, &instr);
|
|
|
|
write_sm4_block(tpf, &iff->then_block);
|
|
|
|
if (!list_empty(&iff->else_block.instrs))
|
|
{
|
|
instr.opcode = VKD3D_SM4_OP_ELSE;
|
|
instr.src_count = 0;
|
|
write_sm4_instruction(tpf, &instr);
|
|
|
|
write_sm4_block(tpf, &iff->else_block);
|
|
}
|
|
|
|
instr.opcode = VKD3D_SM4_OP_ENDIF;
|
|
instr.src_count = 0;
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_jump(const struct tpf_compiler *tpf, const struct hlsl_ir_jump *jump)
|
|
{
|
|
struct sm4_instruction instr = {0};
|
|
|
|
switch (jump->type)
|
|
{
|
|
case HLSL_IR_JUMP_BREAK:
|
|
instr.opcode = VKD3D_SM4_OP_BREAK;
|
|
break;
|
|
|
|
case HLSL_IR_JUMP_CONTINUE:
|
|
instr.opcode = VKD3D_SM4_OP_CONTINUE;
|
|
break;
|
|
|
|
case HLSL_IR_JUMP_DISCARD_NZ:
|
|
{
|
|
instr.opcode = VKD3D_SM4_OP_DISCARD;
|
|
instr.extra_bits = VKD3D_SM4_CONDITIONAL_NZ;
|
|
|
|
memset(&instr.srcs[0], 0, sizeof(*instr.srcs));
|
|
instr.src_count = 1;
|
|
sm4_src_from_node(tpf, &instr.srcs[0], jump->condition.node, VKD3DSP_WRITEMASK_ALL);
|
|
break;
|
|
}
|
|
|
|
case HLSL_IR_JUMP_RETURN:
|
|
vkd3d_unreachable();
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &jump->node.loc, "Jump type %s.", hlsl_jump_type_to_string(jump->type));
|
|
return;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
/* Does this variable's data come directly from the API user, rather than being
|
|
* temporary or from a previous shader stage?
|
|
* I.e. is it a uniform or VS input? */
|
|
static bool var_is_user_input(struct hlsl_ctx *ctx, const struct hlsl_ir_var *var)
|
|
{
|
|
if (var->is_uniform)
|
|
return true;
|
|
|
|
return var->is_input_semantic && ctx->profile->type == VKD3D_SHADER_TYPE_VERTEX;
|
|
}
|
|
|
|
static void write_sm4_load(const struct tpf_compiler *tpf, const struct hlsl_ir_load *load)
|
|
{
|
|
const struct hlsl_type *type = load->node.data_type;
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], &load->node);
|
|
instr.dst_count = 1;
|
|
|
|
VKD3D_ASSERT(hlsl_is_numeric_type(type));
|
|
if (type->e.numeric.type == HLSL_TYPE_BOOL && var_is_user_input(tpf->ctx, load->src.var))
|
|
{
|
|
struct hlsl_constant_value value;
|
|
|
|
/* Uniform bools can be specified as anything, but internal bools always
|
|
* have 0 for false and ~0 for true. Normalize that here. */
|
|
|
|
instr.opcode = VKD3D_SM4_OP_MOVC;
|
|
|
|
sm4_src_from_deref(tpf, &instr.srcs[0], &load->src, instr.dsts[0].write_mask, &instr);
|
|
|
|
memset(&value, 0xff, sizeof(value));
|
|
sm4_src_from_constant_value(&instr.srcs[1], &value, type->dimx, instr.dsts[0].write_mask);
|
|
memset(&value, 0, sizeof(value));
|
|
sm4_src_from_constant_value(&instr.srcs[2], &value, type->dimx, instr.dsts[0].write_mask);
|
|
instr.src_count = 3;
|
|
}
|
|
else
|
|
{
|
|
instr.opcode = VKD3D_SM4_OP_MOV;
|
|
|
|
sm4_src_from_deref(tpf, &instr.srcs[0], &load->src, instr.dsts[0].write_mask, &instr);
|
|
instr.src_count = 1;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_loop(const struct tpf_compiler *tpf, const struct hlsl_ir_loop *loop)
|
|
{
|
|
struct sm4_instruction instr =
|
|
{
|
|
.opcode = VKD3D_SM4_OP_LOOP,
|
|
};
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
|
|
write_sm4_block(tpf, &loop->body);
|
|
|
|
instr.opcode = VKD3D_SM4_OP_ENDLOOP;
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_gather(const struct tpf_compiler *tpf, const struct hlsl_ir_node *dst,
|
|
const struct hlsl_deref *resource, const struct hlsl_deref *sampler,
|
|
const struct hlsl_ir_node *coords, uint32_t swizzle, const struct hlsl_ir_node *texel_offset)
|
|
{
|
|
struct vkd3d_shader_src_param *src;
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
|
|
instr.opcode = VKD3D_SM4_OP_GATHER4;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], dst);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[instr.src_count++], coords, VKD3DSP_WRITEMASK_ALL);
|
|
|
|
if (texel_offset)
|
|
{
|
|
if (!encode_texel_offset_as_aoffimmi(&instr, texel_offset))
|
|
{
|
|
if (tpf->ctx->profile->major_version < 5)
|
|
{
|
|
hlsl_error(tpf->ctx, &texel_offset->loc, VKD3D_SHADER_ERROR_HLSL_INVALID_TEXEL_OFFSET,
|
|
"Offset must resolve to integer literal in the range -8 to 7 for profiles < 5.");
|
|
return;
|
|
}
|
|
instr.opcode = VKD3D_SM5_OP_GATHER4_PO;
|
|
sm4_src_from_node(tpf, &instr.srcs[instr.src_count++], texel_offset, VKD3DSP_WRITEMASK_ALL);
|
|
}
|
|
}
|
|
|
|
sm4_src_from_deref(tpf, &instr.srcs[instr.src_count++], resource, instr.dsts[0].write_mask, &instr);
|
|
|
|
src = &instr.srcs[instr.src_count++];
|
|
sm4_src_from_deref(tpf, src, sampler, VKD3DSP_WRITEMASK_ALL, &instr);
|
|
src->reg.dimension = VSIR_DIMENSION_VEC4;
|
|
src->swizzle = swizzle;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_resource_load(const struct tpf_compiler *tpf, const struct hlsl_ir_resource_load *load)
|
|
{
|
|
const struct hlsl_ir_node *texel_offset = load->texel_offset.node;
|
|
const struct hlsl_ir_node *sample_index = load->sample_index.node;
|
|
const struct hlsl_ir_node *coords = load->coords.node;
|
|
|
|
if (load->sampler.var && !load->sampler.var->is_uniform)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &load->node.loc, "Sample using non-uniform sampler variable.");
|
|
return;
|
|
}
|
|
|
|
if (!load->resource.var->is_uniform)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &load->node.loc, "Load from non-uniform resource variable.");
|
|
return;
|
|
}
|
|
|
|
switch (load->load_type)
|
|
{
|
|
case HLSL_RESOURCE_LOAD:
|
|
write_sm4_ld(tpf, &load->node, &load->resource,
|
|
coords, sample_index, texel_offset, load->sampling_dim);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE:
|
|
case HLSL_RESOURCE_SAMPLE_CMP:
|
|
case HLSL_RESOURCE_SAMPLE_CMP_LZ:
|
|
case HLSL_RESOURCE_SAMPLE_LOD:
|
|
case HLSL_RESOURCE_SAMPLE_LOD_BIAS:
|
|
case HLSL_RESOURCE_SAMPLE_GRAD:
|
|
/* Combined sample expressions were lowered. */
|
|
VKD3D_ASSERT(load->sampler.var);
|
|
write_sm4_sample(tpf, load);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_GATHER_RED:
|
|
write_sm4_gather(tpf, &load->node, &load->resource, &load->sampler, coords,
|
|
VKD3D_SHADER_SWIZZLE(X, X, X, X), texel_offset);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_GATHER_GREEN:
|
|
write_sm4_gather(tpf, &load->node, &load->resource, &load->sampler, coords,
|
|
VKD3D_SHADER_SWIZZLE(Y, Y, Y, Y), texel_offset);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_GATHER_BLUE:
|
|
write_sm4_gather(tpf, &load->node, &load->resource, &load->sampler, coords,
|
|
VKD3D_SHADER_SWIZZLE(Z, Z, Z, Z), texel_offset);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_GATHER_ALPHA:
|
|
write_sm4_gather(tpf, &load->node, &load->resource, &load->sampler, coords,
|
|
VKD3D_SHADER_SWIZZLE(W, W, W, W), texel_offset);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_INFO:
|
|
write_sm4_sampleinfo(tpf, load);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_RESINFO:
|
|
write_sm4_resinfo(tpf, load);
|
|
break;
|
|
|
|
case HLSL_RESOURCE_SAMPLE_PROJ:
|
|
vkd3d_unreachable();
|
|
}
|
|
}
|
|
|
|
static void write_sm4_resource_store(const struct tpf_compiler *tpf, const struct hlsl_ir_resource_store *store)
|
|
{
|
|
struct hlsl_type *resource_type = hlsl_deref_get_type(tpf->ctx, &store->resource);
|
|
|
|
if (!store->resource.var->is_uniform)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &store->node.loc, "Store to non-uniform resource variable.");
|
|
return;
|
|
}
|
|
|
|
if (resource_type->sampler_dim == HLSL_SAMPLER_DIM_STRUCTURED_BUFFER)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &store->node.loc, "Structured buffers store is not implemented.");
|
|
return;
|
|
}
|
|
|
|
write_sm4_store_uav_typed(tpf, &store->resource, store->coords.node, store->value.node);
|
|
}
|
|
|
|
static void write_sm4_store(const struct tpf_compiler *tpf, const struct hlsl_ir_store *store)
|
|
{
|
|
const struct hlsl_ir_node *rhs = store->rhs.node;
|
|
struct sm4_instruction instr;
|
|
uint32_t writemask;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_MOV;
|
|
|
|
sm4_register_from_deref(tpf->ctx, &instr.dsts[0].reg, &writemask, &store->lhs, &instr);
|
|
instr.dsts[0].write_mask = hlsl_combine_writemasks(writemask, store->writemask);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], rhs, instr.dsts[0].write_mask);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_switch(const struct tpf_compiler *tpf, const struct hlsl_ir_switch *s)
|
|
{
|
|
const struct hlsl_ir_node *selector = s->selector.node;
|
|
struct hlsl_ir_switch_case *c;
|
|
struct sm4_instruction instr;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_SWITCH;
|
|
|
|
sm4_src_from_node(tpf, &instr.srcs[0], selector, VKD3DSP_WRITEMASK_ALL);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
|
|
LIST_FOR_EACH_ENTRY(c, &s->cases, struct hlsl_ir_switch_case, entry)
|
|
{
|
|
memset(&instr, 0, sizeof(instr));
|
|
if (c->is_default)
|
|
{
|
|
instr.opcode = VKD3D_SM4_OP_DEFAULT;
|
|
}
|
|
else
|
|
{
|
|
struct hlsl_constant_value value = { .u[0].u = c->value };
|
|
|
|
instr.opcode = VKD3D_SM4_OP_CASE;
|
|
sm4_src_from_constant_value(&instr.srcs[0], &value, 1, VKD3DSP_WRITEMASK_ALL);
|
|
instr.src_count = 1;
|
|
}
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
write_sm4_block(tpf, &c->body);
|
|
}
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_ENDSWITCH;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_swizzle(const struct tpf_compiler *tpf, const struct hlsl_ir_swizzle *swizzle)
|
|
{
|
|
unsigned int hlsl_swizzle;
|
|
struct sm4_instruction instr;
|
|
uint32_t writemask;
|
|
|
|
memset(&instr, 0, sizeof(instr));
|
|
instr.opcode = VKD3D_SM4_OP_MOV;
|
|
|
|
sm4_dst_from_node(&instr.dsts[0], &swizzle->node);
|
|
instr.dst_count = 1;
|
|
|
|
sm4_register_from_node(&instr.srcs[0].reg, &writemask, swizzle->val.node);
|
|
hlsl_swizzle = hlsl_map_swizzle(hlsl_combine_swizzles(hlsl_swizzle_from_writemask(writemask),
|
|
swizzle->swizzle, swizzle->node.data_type->dimx), instr.dsts[0].write_mask);
|
|
instr.srcs[0].swizzle = swizzle_from_sm4(hlsl_swizzle);
|
|
instr.src_count = 1;
|
|
|
|
write_sm4_instruction(tpf, &instr);
|
|
}
|
|
|
|
static void write_sm4_block(const struct tpf_compiler *tpf, const struct hlsl_block *block)
|
|
{
|
|
const struct hlsl_ir_node *instr;
|
|
|
|
LIST_FOR_EACH_ENTRY(instr, &block->instrs, struct hlsl_ir_node, entry)
|
|
{
|
|
if (instr->data_type)
|
|
{
|
|
if (instr->data_type->class != HLSL_CLASS_SCALAR && instr->data_type->class != HLSL_CLASS_VECTOR)
|
|
{
|
|
hlsl_fixme(tpf->ctx, &instr->loc, "Class %#x should have been lowered or removed.",
|
|
instr->data_type->class);
|
|
break;
|
|
}
|
|
|
|
if (!instr->reg.allocated)
|
|
{
|
|
VKD3D_ASSERT(instr->type == HLSL_IR_CONSTANT);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
switch (instr->type)
|
|
{
|
|
case HLSL_IR_CALL:
|
|
case HLSL_IR_CONSTANT:
|
|
vkd3d_unreachable();
|
|
|
|
case HLSL_IR_EXPR:
|
|
write_sm4_expr(tpf, hlsl_ir_expr(instr));
|
|
break;
|
|
|
|
case HLSL_IR_IF:
|
|
write_sm4_if(tpf, hlsl_ir_if(instr));
|
|
break;
|
|
|
|
case HLSL_IR_JUMP:
|
|
write_sm4_jump(tpf, hlsl_ir_jump(instr));
|
|
break;
|
|
|
|
case HLSL_IR_LOAD:
|
|
write_sm4_load(tpf, hlsl_ir_load(instr));
|
|
break;
|
|
|
|
case HLSL_IR_RESOURCE_LOAD:
|
|
write_sm4_resource_load(tpf, hlsl_ir_resource_load(instr));
|
|
break;
|
|
|
|
case HLSL_IR_RESOURCE_STORE:
|
|
write_sm4_resource_store(tpf, hlsl_ir_resource_store(instr));
|
|
break;
|
|
|
|
case HLSL_IR_LOOP:
|
|
write_sm4_loop(tpf, hlsl_ir_loop(instr));
|
|
break;
|
|
|
|
case HLSL_IR_STORE:
|
|
write_sm4_store(tpf, hlsl_ir_store(instr));
|
|
break;
|
|
|
|
case HLSL_IR_SWITCH:
|
|
write_sm4_switch(tpf, hlsl_ir_switch(instr));
|
|
break;
|
|
|
|
case HLSL_IR_SWIZZLE:
|
|
write_sm4_swizzle(tpf, hlsl_ir_swizzle(instr));
|
|
break;
|
|
|
|
default:
|
|
hlsl_fixme(tpf->ctx, &instr->loc, "Instruction type %s.", hlsl_node_type_to_string(instr->type));
|
|
}
|
|
}
|
|
}
|
|
|
|
static void write_sm4_shdr(struct hlsl_ctx *ctx, struct hlsl_ir_function_decl *entry_func,
|
|
struct sm4_stat *stat, struct dxbc_writer *dxbc)
|
|
{
|
|
const struct hlsl_profile_info *profile = ctx->profile;
|
|
struct vkd3d_bytecode_buffer buffer = {0};
|
|
struct extern_resource *extern_resources;
|
|
unsigned int extern_resources_count, i;
|
|
const struct hlsl_buffer *cbuffer;
|
|
const struct hlsl_scope *scope;
|
|
const struct hlsl_ir_var *var;
|
|
size_t token_count_position;
|
|
struct tpf_compiler tpf;
|
|
uint32_t temp_count;
|
|
|
|
static const uint16_t shader_types[VKD3D_SHADER_TYPE_COUNT] =
|
|
{
|
|
VKD3D_SM4_PS,
|
|
VKD3D_SM4_VS,
|
|
VKD3D_SM4_GS,
|
|
VKD3D_SM5_HS,
|
|
VKD3D_SM5_DS,
|
|
VKD3D_SM5_CS,
|
|
0, /* EFFECT */
|
|
0, /* TEXTURE */
|
|
VKD3D_SM4_LIB,
|
|
};
|
|
|
|
mark_indexable_vars(ctx, entry_func);
|
|
temp_count = allocate_temp_registers(ctx, entry_func);
|
|
if (ctx->result)
|
|
return;
|
|
|
|
tpf_compiler_init(&tpf, ctx, stat, &buffer);
|
|
|
|
extern_resources = sm4_get_extern_resources(ctx, &extern_resources_count);
|
|
|
|
put_u32(&buffer, vkd3d_make_u32((profile->major_version << 4) | profile->minor_version, shader_types[profile->type]));
|
|
token_count_position = put_u32(&buffer, 0);
|
|
|
|
LIST_FOR_EACH_ENTRY(cbuffer, &ctx->buffers, struct hlsl_buffer, entry)
|
|
{
|
|
if (cbuffer->reg.allocated)
|
|
write_sm4_dcl_constant_buffer(&tpf, cbuffer);
|
|
}
|
|
|
|
for (i = 0; i < extern_resources_count; ++i)
|
|
{
|
|
const struct extern_resource *resource = &extern_resources[i];
|
|
|
|
if (resource->regset == HLSL_REGSET_SAMPLERS)
|
|
write_sm4_dcl_samplers(&tpf, resource);
|
|
else if (resource->regset == HLSL_REGSET_TEXTURES)
|
|
write_sm4_dcl_textures(&tpf, resource, false);
|
|
else if (resource->regset == HLSL_REGSET_UAVS)
|
|
write_sm4_dcl_textures(&tpf, resource, true);
|
|
}
|
|
|
|
if (entry_func->early_depth_test && profile->major_version >= 5)
|
|
write_sm4_dcl_global_flags(&tpf, VKD3DSGF_FORCE_EARLY_DEPTH_STENCIL);
|
|
|
|
LIST_FOR_EACH_ENTRY(var, &entry_func->extern_vars, struct hlsl_ir_var, extern_entry)
|
|
{
|
|
if ((var->is_input_semantic && var->last_read) || (var->is_output_semantic && var->first_write))
|
|
write_sm4_dcl_semantic(&tpf, var);
|
|
}
|
|
|
|
if (profile->type == VKD3D_SHADER_TYPE_COMPUTE)
|
|
write_sm4_dcl_thread_group(&tpf, ctx->thread_count);
|
|
|
|
if (temp_count)
|
|
write_sm4_dcl_temps(&tpf, temp_count);
|
|
|
|
LIST_FOR_EACH_ENTRY(scope, &ctx->scopes, struct hlsl_scope, entry)
|
|
{
|
|
LIST_FOR_EACH_ENTRY(var, &scope->vars, struct hlsl_ir_var, scope_entry)
|
|
{
|
|
if (var->is_uniform || var->is_input_semantic || var->is_output_semantic)
|
|
continue;
|
|
if (!var->regs[HLSL_REGSET_NUMERIC].allocated)
|
|
continue;
|
|
|
|
if (var->indexable)
|
|
{
|
|
unsigned int id = var->regs[HLSL_REGSET_NUMERIC].id;
|
|
unsigned int size = align(var->data_type->reg_size[HLSL_REGSET_NUMERIC], 4) / 4;
|
|
|
|
write_sm4_dcl_indexable_temp(&tpf, id, size, 4);
|
|
}
|
|
}
|
|
}
|
|
|
|
write_sm4_block(&tpf, &entry_func->body);
|
|
|
|
write_sm4_ret(&tpf);
|
|
|
|
set_u32(&buffer, token_count_position, bytecode_get_size(&buffer) / sizeof(uint32_t));
|
|
|
|
add_section(ctx, dxbc, TAG_SHDR, &buffer);
|
|
|
|
sm4_free_extern_resources(extern_resources, extern_resources_count);
|
|
}
|
|
|
|
static void write_sm4_sfi0(struct hlsl_ctx *ctx, struct dxbc_writer *dxbc)
|
|
{
|
|
struct extern_resource *extern_resources;
|
|
unsigned int extern_resources_count;
|
|
uint64_t *flags;
|
|
|
|
flags = vkd3d_calloc(1, sizeof(*flags));
|
|
|
|
extern_resources = sm4_get_extern_resources(ctx, &extern_resources_count);
|
|
for (unsigned int i = 0; i < extern_resources_count; ++i)
|
|
{
|
|
if (extern_resources[i].component_type && extern_resources[i].component_type->e.resource.rasteriser_ordered)
|
|
*flags |= VKD3D_SM4_REQUIRES_ROVS;
|
|
}
|
|
sm4_free_extern_resources(extern_resources, extern_resources_count);
|
|
|
|
/* FIXME: We also emit code that should require UAVS_AT_EVERY_STAGE,
|
|
* STENCIL_REF, and TYPED_UAV_LOAD_ADDITIONAL_FORMATS. */
|
|
|
|
if (*flags)
|
|
dxbc_writer_add_section(dxbc, TAG_SFI0, flags, sizeof(*flags));
|
|
else
|
|
vkd3d_free(flags);
|
|
}
|
|
|
|
static void write_sm4_stat(struct hlsl_ctx *ctx, const struct sm4_stat *stat, struct dxbc_writer *dxbc)
|
|
{
|
|
struct vkd3d_bytecode_buffer buffer = {0};
|
|
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_INSTR_COUNT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_TEMPS]);
|
|
put_u32(&buffer, 0); /* Def count */
|
|
put_u32(&buffer, 0); /* DCL count */
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_FLOAT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_INT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_UINT]);
|
|
put_u32(&buffer, 0); /* Static flow control count */
|
|
put_u32(&buffer, 0); /* Dynamic flow control count */
|
|
put_u32(&buffer, 0); /* Macro instruction count */
|
|
put_u32(&buffer, 0); /* Temp array count */
|
|
put_u32(&buffer, 0); /* Array instr count */
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_CUT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_EMIT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_SAMPLE]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_LOAD]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_SAMPLE_C]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_SAMPLE_BIAS]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_SAMPLE_GRAD]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_MOV]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_MOVC]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_CONV]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_BITWISE]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_DCL_INPUT_PRIMITIVE]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_DCL_OUTPUT_TOPOLOGY]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_DCL_VERTICES_OUT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_GATHER]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_LOD]);
|
|
put_u32(&buffer, 0); /* Sample frequency */
|
|
|
|
if (hlsl_version_ge(ctx, 5, 0))
|
|
{
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_DCL_GS_INSTANCES]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_TESS_CONTROL_POINT_COUNT]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_TESS_OUTPUT_PRIMITIVE]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_TESS_PARTITIONING]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_TESS_DOMAIN]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_BARRIER]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_ATOMIC]);
|
|
put_u32(&buffer, stat->fields[VKD3D_STAT_STORE]);
|
|
}
|
|
|
|
add_section(ctx, dxbc, TAG_STAT, &buffer);
|
|
}
|
|
|
|
/* OBJECTIVE: Stop relying on ctx and entry_func on this function, receiving
|
|
* data from the other parameters instead, so they can be removed from the
|
|
* arguments and this function can be independent of HLSL structs. */
|
|
int tpf_compile(struct vsir_program *program, uint64_t config_flags,
|
|
struct vkd3d_shader_code *out, struct vkd3d_shader_message_context *message_context,
|
|
struct hlsl_ctx *ctx, struct hlsl_ir_function_decl *entry_func)
|
|
{
|
|
struct sm4_stat stat = {0};
|
|
struct dxbc_writer dxbc;
|
|
size_t i;
|
|
int ret;
|
|
|
|
dxbc_writer_init(&dxbc);
|
|
|
|
write_sm4_signature(ctx, &dxbc, entry_func, false);
|
|
write_sm4_signature(ctx, &dxbc, entry_func, true);
|
|
write_sm4_rdef(ctx, &dxbc);
|
|
write_sm4_shdr(ctx, entry_func, &stat, &dxbc);
|
|
write_sm4_sfi0(ctx, &dxbc);
|
|
write_sm4_stat(ctx, &stat, &dxbc);
|
|
|
|
if (!(ret = ctx->result))
|
|
ret = dxbc_writer_write(&dxbc, out);
|
|
for (i = 0; i < dxbc.section_count; ++i)
|
|
vkd3d_shader_free_shader_code(&dxbc.sections[i].data);
|
|
return ret;
|
|
}
|